sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
sequencelengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
sequencelengths 0
201
| languages
sequencelengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
sequencelengths 0
722
| processed_texts
sequencelengths 1
723
| tokens_length
sequencelengths 1
723
| input_texts
sequencelengths 1
61
| embeddings
sequencelengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null | null |
<!-- header start -->
<!-- 200823 -->
<div style="width: auto; margin-left: auto; margin-right: auto">
<img src="https://github.com/second-state/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
</div>
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
<!-- header end -->
# Qwen1.5-1.8B-Chat-GGUF
## Original Model
[Qwen/Qwen1.5-1.8B-Chat](https://huggingface.co/Qwen/Qwen1.5-1.8B-Chat)
## Run with LlamaEdge
- LlamaEdge version: [v0.2.15](https://github.com/second-state/LlamaEdge/releases/tag/0.2.15) and above
- Prompt template
- Prompt type: `chatml`
- Prompt string
```text
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
```
- Run as LlamaEdge service
```bash
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Qwen1.5-1.8B-Chat-Q5_K_M.gguf llama-api-server.wasm -p chatml
```
- Run as LlamaEdge command app
```bash
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Qwen1.5-1.8B-Chat-Q5_K_M.gguf llama-chat.wasm -p chatml
```
## Quantized GGUF Models
| Name | Quant method | Bits | Size | Use case |
| ---- | ---- | ---- | ---- | ----- |
| [Qwen1.5-1.8B-Chat-Q2_K.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q2_K.gguf) | Q2_K | 2 | 863 MB| smallest, significant quality loss - not recommended for most purposes |
| [Qwen1.5-1.8B-Chat-Q3_K_L.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q3_K_L.gguf) | Q3_K_L | 3 | 1.06 GB| small, substantial quality loss |
| [Qwen1.5-1.8B-Chat-Q3_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q3_K_M.gguf) | Q3_K_M | 3 | 1.02 GB| very small, high quality loss |
| [Qwen1.5-1.8B-Chat-Q3_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q3_K_S.gguf) | Q3_K_S | 3 | 970 MB| very small, high quality loss |
| [Qwen1.5-1.8B-Chat-Q4_0.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q4_0.gguf) | Q4_0 | 4 | 1.12 GB| legacy; small, very high quality loss - prefer using Q3_K_M |
| [Qwen1.5-1.8B-Chat-Q4_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q4_K_M.gguf) | Q4_K_M | 4 | 1.22 GB| medium, balanced quality - recommended |
| [Qwen1.5-1.8B-Chat-Q4_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q4_K_S.gguf) | Q4_K_S | 4 | 1.16 GB| small, greater quality loss |
| [Qwen1.5-1.8B-Chat-Q5_0.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q5_0.gguf) | Q5_0 | 5 | 1.31 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
| [Qwen1.5-1.8B-Chat-Q5_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q5_K_M.gguf) | Q5_K_M | 5 | 1.38 GB| large, very low quality loss - recommended |
| [Qwen1.5-1.8B-Chat-Q5_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q5_K_S.gguf) | Q5_K_S | 5 | 1.33 GB| large, low quality loss - recommended |
| [Qwen1.5-1.8B-Chat-Q6_K.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q6_K.gguf) | Q6_K | 6 | 1.58 GB| very large, extremely low quality loss |
| [Qwen1.5-1.8B-Chat-Q8_0.gguf](https://huggingface.co/second-state/Qwen1.5-1.8B-Chat-GGUF/blob/main/Qwen1.5-1.8B-Chat-Q8_0.gguf) | Q8_0 | 8 | 1.96 GB| very large, extremely low quality loss - not recommended |
| {"language": ["en"], "license": "other", "tags": ["chat"], "model_name": "Qwen1.5 1.8B Chat", "base_model": "Qwen/Qwen1.5-1.8B-Chat", "license_name": "tongyi-qianwen-research", "license_link": "https://huggingface.co/Qwen/Qwen1.5-1.8B-Chat/blob/main/LICENSE", "model_creator": "Qwen", "quantized_by": "Second State Inc.", "pipeline_tag": "text-generation"} | text-generation | second-state/Qwen1.5-1.8B-Chat-GGUF | [
"gguf",
"chat",
"text-generation",
"en",
"base_model:Qwen/Qwen1.5-1.8B-Chat",
"license:other",
"region:us"
] | 2024-02-06T04:33:23+00:00 | [] | [
"en"
] | TAGS
#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-1.8B-Chat #license-other #region-us
|

---
Qwen1.5-1.8B-Chat-GGUF
======================
Original Model
--------------
Qwen/Qwen1.5-1.8B-Chat
Run with LlamaEdge
------------------
* LlamaEdge version: v0.2.15 and above
* Prompt template
+ Prompt type: 'chatml'
+ Prompt string
* Run as LlamaEdge service
* Run as LlamaEdge command app
Quantized GGUF Models
---------------------
| [] | [
"TAGS\n#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-1.8B-Chat #license-other #region-us \n"
] | [
39
] | [
"passage: TAGS\n#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-1.8B-Chat #license-other #region-us \n"
] | [
-0.012159162200987339,
-0.002384505234658718,
-0.0036800289526581764,
-0.011382674798369408,
0.01817602664232254,
0.0013886467786505818,
0.19365325570106506,
0.0920649990439415,
0.20770816504955292,
-0.054096776992082596,
0.0969669446349144,
-0.046702586114406586,
0.0560922771692276,
0.10536717623472214,
0.03941493481397629,
-0.1732979714870453,
0.06661741435527802,
-0.04739951714873314,
-0.0061686052940785885,
0.05550824850797653,
0.06539178639650345,
0.021326808258891106,
0.09493700414896011,
-0.001204302883706987,
-0.09429270774126053,
0.02829204685986042,
-0.009697887115180492,
-0.03361004590988159,
0.08377794921398163,
0.10066759586334229,
-0.015915056690573692,
0.10455561429262161,
-0.046870529651641846,
-0.2266409695148468,
0.04012414440512657,
-0.04650779068470001,
-0.1433538794517517,
-0.0065388306975364685,
0.002318168058991432,
0.027921244502067566,
0.076313816010952,
0.1909319907426834,
-0.08731281012296677,
0.13320577144622803,
-0.12949036061763763,
-0.06624457985162735,
-0.07773614674806595,
0.09555815905332565,
0.016280822455883026,
0.05452343076467514,
-0.027829736471176147,
0.043647024780511856,
-0.11306392401456833,
0.02278152108192444,
0.10090848803520203,
-0.4887355864048004,
0.05517465993762016,
0.2765573561191559,
0.11026635020971298,
0.13922949135303497,
-0.07605817914009094,
0.042336881160736084,
0.05801092088222504,
-0.011525226756930351,
-0.18309658765792847,
-0.023169850930571556,
0.06529004871845245,
0.061002496629953384,
-0.04943687841296196,
-0.0101597486063838,
0.2459317147731781,
0.07998159527778625,
-0.009039243683218956,
0.023593932390213013,
-0.015546971932053566,
-0.030315494164824486,
-0.023540707305073738,
0.06832093745470047,
-0.04725434631109238,
0.17478372156620026,
-0.00006089619273552671,
-0.09035971015691757,
-0.1065218523144722,
-0.055883340537548065,
-0.17722347378730774,
0.005147211253643036,
0.008668407797813416,
0.09591832011938095,
-0.1734026074409485,
0.05523096024990082,
-0.16459950804710388,
-0.07686230540275574,
-0.08173322677612305,
-0.1276068538427353,
0.025183655321598053,
0.07656692713499069,
0.003442061133682728,
0.12121708691120148,
0.21212074160575867,
0.14165887236595154,
-0.08298340439796448,
0.07351001352071762,
-0.06083006039261818,
0.10268956422805786,
0.03140123933553696,
0.0015262254746630788,
0.05487845093011856,
-0.02269524335861206,
0.048548437654972076,
-0.09027034044265747,
-0.020168805494904518,
-0.0459260568022728,
-0.1711028814315796,
0.0006761599215678871,
-0.12031232565641403,
0.07901086658239365,
0.005289100576192141,
0.018791450187563896,
0.012456297874450684,
-0.0014547844184562564,
0.22857728600502014,
0.04787556082010269,
-0.06254248321056366,
0.05812903866171837,
-0.008618216030299664,
-0.09517502039670944,
-0.012850531376898289,
0.09669660031795502,
0.05815249681472778,
-0.14199957251548767,
-0.05388639122247696,
-0.0900634303689003,
0.036016955971717834,
0.03428530693054199,
0.03211897611618042,
0.047227393835783005,
0.018127350136637688,
-0.1449964940547943,
-0.08495248854160309,
0.03878294676542282,
0.003054940840229392,
-0.03148473799228668,
-0.08562494069337845,
-0.05148039758205414,
-0.048867713660001755,
0.0006323695997707546,
-0.05916077643632889,
-0.002792198909446597,
-0.09321975708007812,
0.07577717304229736,
-0.04598364606499672,
0.008355182595551014,
-0.2623205780982971,
0.005010803230106831,
-0.02728419192135334,
0.0202485341578722,
-0.061726026237010956,
0.10737426578998566,
-0.18229040503501892,
0.08675925433635712,
-0.02306337095797062,
0.04360917583107948,
-0.08604514598846436,
0.010728859342634678,
-0.06802890449762344,
0.15414945781230927,
-0.10776004940271378,
-0.08992833644151688,
0.2802131175994873,
-0.10355233401060104,
-0.1296473890542984,
0.09681108593940735,
-0.0001802056358428672,
-0.03359582647681236,
0.046320050954818726,
0.3852606415748596,
0.06101797893643379,
-0.06582729518413544,
-0.029801929369568825,
0.18243518471717834,
-0.10279884189367294,
-0.07841824740171432,
0.0886794701218605,
-0.07605163007974625,
-0.08132325857877731,
0.04265746474266052,
0.019174925982952118,
0.13832364976406097,
-0.05245092138648033,
-0.06849502772092819,
-0.03515800088644028,
-0.06252452731132507,
0.015591088682413101,
-0.03933000937104225,
0.0658867359161377,
-0.058151066303253174,
-0.020640095695853233,
-0.11468856036663055,
0.08423053473234177,
0.07850957661867142,
-0.026607826352119446,
-0.13425986468791962,
0.06073218584060669,
0.1017138659954071,
0.05013227462768555,
-0.0014180588768795133,
-0.09390313923358917,
0.018323948606848717,
0.05222329497337341,
0.09559450298547745,
0.12371313571929932,
0.051626238971948624,
-0.030173365026712418,
0.0010515042813494802,
-0.005328419618308544,
0.029468359425663948,
-0.011110746301710606,
-0.03592490777373314,
-0.09672782570123672,
0.03219108656048775,
-0.04305353760719299,
0.11832349002361298,
-0.08226782828569412,
0.0029775514267385006,
0.0967431291937828,
0.0430268794298172,
-0.056258756667375565,
0.03469231352210045,
-0.016841625794768333,
-0.0403701476752758,
0.0021904222667217255,
-0.006769325118511915,
0.06994596868753433,
0.03654734417796135,
-0.12874512374401093,
0.22967825829982758,
-0.044466283172369,
0.08958127349615097,
0.12141053378582001,
-0.010613223537802696,
0.07785595953464508,
-0.12011432647705078,
-0.06213178113102913,
-0.006777183152735233,
0.10572244971990585,
-0.01948263682425022,
0.2283279299736023,
-0.04664591699838638,
0.08127415925264359,
-0.07028853893280029,
-0.03923128917813301,
-0.012442785315215588,
0.008137662895023823,
-0.019901061430573463,
0.09995247423648834,
0.09131903201341629,
-0.18384546041488647,
0.0953700914978981,
0.1615479588508606,
0.22310051321983337,
0.26917359232902527,
-0.0904335230588913,
0.04811258241534233,
-0.011514155194163322,
0.06456723064184189,
-0.08372269570827484,
0.11818879097700119,
-0.21636293828487396,
-0.015468785539269447,
0.02707725577056408,
0.05929669737815857,
0.135161355137825,
-0.06629273295402527,
-0.11122652888298035,
-0.053410232067108154,
-0.1071772500872612,
-0.14142172038555145,
0.09756964445114136,
-0.034169211983680725,
0.055769920349121094,
0.017297381535172462,
-0.025051424279808998,
0.07778181880712509,
-0.03829789534211159,
-0.07569608092308044,
0.13446548581123352,
-0.1513601690530777,
-0.16461703181266785,
-0.04993819072842598,
-0.06701960414648056,
-0.09500487893819809,
-0.002522359136492014,
0.10791829228401184,
-0.17736360430717468,
0.014980804175138474,
0.005328523460775614,
-0.028912130743265152,
-0.09612986445426941,
0.007652322296053171,
0.021766595542430878,
-0.015074891038239002,
-0.04040602594614029,
-0.15713372826576233,
-0.03818964958190918,
-0.034586090594530106,
-0.12562954425811768,
0.09236907213926315,
-0.1666952222585678,
0.011388052254915237,
0.13328295946121216,
0.050116945058107376,
0.08878358453512192,
-0.04984689876437187,
0.3436707556247711,
-0.04916393384337425,
-0.0068149142898619175,
0.06885436922311783,
0.08989690989255905,
0.05648117884993553,
0.12260445952415466,
0.02415713109076023,
-0.09415487945079803,
0.0036502808798104525,
0.017988556995987892,
-0.06892988085746765,
-0.18353553116321564,
-0.1087692379951477,
-0.02635539323091507,
0.15881864726543427,
-0.09505215287208557,
0.1511709988117218,
0.10435117781162262,
-0.01647144928574562,
0.007578406948596239,
-0.06893017143011093,
0.06111711263656616,
0.030804337933659554,
0.03280167654156685,
-0.053735457360744476,
0.06491547077894211,
-0.06183159723877907,
0.07397160679101944,
0.1434783786535263,
0.11832563579082489,
0.11174560338258743,
0.1342105120420456,
0.07877985388040543,
0.11731452494859695,
0.267412930727005,
0.19269898533821106,
-0.1512422412633896,
-0.012741154991090298,
-0.08236320316791534,
-0.03257070481777191,
-0.024987170472741127,
0.024663129821419716,
0.010224875062704086,
-0.006941009312868118,
-0.19830262660980225,
-0.039601631462574005,
-0.05685042217373848,
0.03787922486662865,
0.01622815802693367,
0.03545496612787247,
0.08978688716888428,
-0.004706461913883686,
0.04625049978494644,
-0.02538963221013546,
-0.005201447755098343,
0.14251533150672913,
0.08568192273378372,
-0.0998254269361496,
0.02477257512509823,
0.030400870367884636,
0.10516943037509918,
0.06022333726286888,
0.04755634069442749,
-0.07448923587799072,
-0.07069365680217743,
0.006026772782206535,
0.11402937769889832,
-0.19906839728355408,
0.2045307606458664,
0.0018004601588472724,
-0.04304938763380051,
-0.033617619425058365,
-0.07640237361192703,
0.012712005525827408,
0.12505917251110077,
0.0858493372797966,
0.07289421558380127,
-0.08799894154071808,
0.025806471705436707,
-0.06445170938968658,
0.1020529568195343,
0.07214392721652985,
-0.021904146298766136,
-0.08796263486146927,
-0.0005037077353335917,
0.026789210736751556,
-0.006695255171507597,
0.05038398876786232,
-0.14809243381023407,
-0.09949895739555359,
-0.021669115871191025,
0.09601256251335144,
0.09912284463644028,
-0.06425812840461731,
0.06415010243654251,
-0.07775092869997025,
0.08517636358737946,
-0.08491096645593643,
-0.07164017856121063,
-0.0405711904168129,
-0.0712958499789238,
0.01689852960407734,
-0.051262859255075455,
-0.016010092571377754,
-0.05687025561928749,
-0.01884809136390686,
-0.10279318690299988,
-0.21364153921604156,
0.0613955557346344,
-0.08025044947862625,
-0.049922917038202286,
-0.010181779973208904,
0.197548970580101,
0.037200238555669785,
0.00646867323666811,
0.041639067232608795,
-0.10120652616024017,
-0.07544854283332825,
-0.19301468133926392,
0.045875534415245056,
-0.10861585289239883,
-0.12403812259435654,
0.00830639898777008,
0.017072100192308426,
-0.010358267463743687,
-0.030845824629068375,
-0.06535526365041733,
0.16886167228221893,
0.3025183081626892,
-0.06111583486199379,
0.21247036755084991,
0.26488760113716125,
-0.054350387305021286,
-0.13722297549247742,
-0.19616326689720154,
-0.08583659678697586,
-0.08297096192836761,
0.015117705799639225,
-0.1696438491344452,
-0.016232723370194435,
0.06637117266654968,
-0.05369070544838905,
0.15151438117027283,
-0.2830822467803955,
-0.037261199206113815,
0.08925077319145203,
-0.07948710769414902,
0.42429742217063904,
-0.13897928595542908,
-0.17259065806865692,
0.007195821963250637,
-0.26590949296951294,
0.14215566217899323,
-0.12960033118724823,
0.09995017200708389,
0.01726062223315239,
0.0898573026061058,
-0.038148581981658936,
0.01130371168255806,
0.134153351187706,
0.046472229063510895,
0.006229916121810675,
-0.08428948372602463,
-0.05786577984690666,
0.06457347422838211,
0.03871795907616615,
-0.032301247119903564,
-0.1821502149105072,
0.05683845654129982,
-0.1503063291311264,
0.013768785633146763,
-0.12000183761119843,
0.025651782751083374,
-0.00637460732832551,
-0.05825076997280121,
-0.09248556941747665,
0.0224488265812397,
-0.04827360063791275,
0.0313827209174633,
0.11060602962970734,
-0.07873306423425674,
0.12221892923116684,
0.1273147612810135,
-0.040801163762807846,
-0.20463921129703522,
-0.004640798084437847,
-0.10683640837669373,
-0.0589129813015461,
0.07066553086042404,
-0.19105692207813263,
-0.03311609849333763,
0.08958374708890915,
-0.03367030620574951,
0.07965380698442459,
0.05038898065686226,
-0.10563027858734131,
0.12456993758678436,
0.12035232037305832,
-0.13553041219711304,
-0.22015126049518585,
-0.06978951394557953,
0.10453487932682037,
0.16878022253513336,
0.07238104194402695,
0.10248322039842606,
0.026011325418949127,
-0.015969695523381233,
0.01918545551598072,
-0.004970915149897337,
-0.11337754875421524,
-0.051901839673519135,
0.05109477415680885,
0.014373836107552052,
-0.15439572930335999,
0.1135278269648552,
0.015135730616748333,
0.10703767836093903,
0.01840297132730484,
0.1614304482936859,
-0.14280147850513458,
-0.08881879597902298,
-0.1878054440021515,
0.14723917841911316,
-0.1475381702184677,
-0.032754525542259216,
0.011536415666341782,
-0.0848340094089508,
0.0031818512361496687,
0.10490988940000534,
-0.02423297055065632,
0.09137299656867981,
0.09741409122943878,
0.017763178795576096,
0.11225931346416473,
-0.04538079351186752,
-0.13416224718093872,
-0.009766818024218082,
-0.0975232794880867,
-0.08077148348093033,
-0.0032085187267512083,
0.15690608322620392,
-0.037512172013521194,
-0.11847541481256485,
-0.1858169585466385,
0.0013745066244155169,
-0.1845809519290924,
-0.03408735245466232,
-0.06361877173185349,
-0.009473612532019615,
-0.013043335638940334,
-0.10599052160978317,
-0.025707818567752838,
0.0058595468290150166,
-0.100814089179039,
-0.027218887582421303,
0.023757167160511017,
0.0689331591129303,
-0.09046777337789536,
0.02579391747713089,
0.07520223408937454,
0.06951872259378433,
0.17219294607639313,
0.12710347771644592,
-0.020250866189599037,
0.13565672934055328,
-0.15070419013500214,
-0.035092588514089584,
0.008817087858915329,
0.004233836196362972,
0.03598010912537575,
0.15654154121875763,
-0.038157716393470764,
-0.012577703222632408,
-0.018871797248721123,
0.05183091759681702,
0.03585270419716835,
-0.12260597944259644,
-0.10208325833082199,
-0.023776225745677948,
-0.027034388855099678,
-0.021972203627228737,
-0.07181980460882187,
0.13685590028762817,
0.06449522823095322,
0.0602986142039299,
-0.05322536081075668,
-0.00318137533031404,
-0.08068037778139114,
0.005544246640056372,
0.042688023298978806,
-0.036287639290094376,
0.007997720502316952,
-0.07678815722465515,
-0.07179150730371475,
-0.024143606424331665,
0.2189880907535553,
-0.06290798634290695,
-0.06500154733657837,
0.026831360533833504,
0.003883258206769824,
0.03268928453326225,
-0.040502652525901794,
0.28498515486717224,
0.08051709085702896,
0.00409980583935976,
-0.0550442673265934,
0.04051506146788597,
-0.0275528896600008,
-0.16514751315116882,
0.025980466976761818,
-0.000575818179640919,
0.02153264358639717,
0.017970476299524307,
0.1821213811635971,
-0.06993091106414795,
-0.006416624411940575,
-0.03562387824058533,
-0.00829380378127098,
-0.014947919175028801,
-0.0840383842587471,
0.06057005003094673,
0.17309711873531342,
-0.034945711493492126,
0.03652114421129227,
0.012529348023235798,
-0.025734150782227516,
-0.06330371648073196,
-0.08271461725234985,
-0.0326492078602314,
-0.22448548674583435,
0.07836873084306717,
-0.0874384269118309,
0.07672767341136932,
0.0730319619178772,
0.05209546163678169,
-0.051811378449201584,
-0.018667234107851982,
-0.028683405369520187,
-0.10334870964288712,
0.028583107516169548,
-0.043845370411872864,
0.0011186482151970267,
-0.08062968403100967,
-0.0033948372583836317,
0.05933539196848869,
-0.005411463789641857,
-0.09601668268442154,
0.08078178018331528,
-0.011861255392432213,
0.00416677026078105,
-0.09847098588943481,
-0.04320213198661804,
-0.04848392680287361,
0.05417049676179886,
0.013465728610754013,
0.20661045610904694,
0.027523528784513474,
0.00717198895290494,
0.04903257638216019,
0.21844497323036194,
0.016394559293985367,
-0.10080061107873917,
-0.03649644926190376,
-0.04953703284263611,
-0.07568259537220001,
0.09856807440519333,
-0.06798958778381348,
-0.009323895908892155,
-0.08254951238632202,
0.26773184537887573,
0.24064308404922485,
-0.11411507427692413,
0.017148887738585472,
-0.03202279657125473,
0.021075190976262093,
0.06897581368684769,
0.11942893266677856,
0.06723964214324951,
0.24534033238887787,
-0.09658895432949066,
-0.017123354598879814,
0.014333049766719341,
0.03235993534326553,
-0.10409832000732422,
0.09206874668598175,
0.008565275929868221,
-0.068283811211586,
-0.060164835304021835,
0.08703646808862686,
-0.2668989598751068,
0.09211305528879166,
-0.1495690643787384,
-0.11434464156627655,
-0.018727684393525124,
-0.08373859524726868,
0.08697710186243057,
0.09118811041116714,
0.05260276421904564,
-0.06500503420829773,
0.006656364072114229,
0.0416816882789135,
-0.010291139595210552,
-0.23701009154319763,
-0.13660094141960144,
0.07783825695514679,
0.014818452298641205,
0.12615546584129333,
-0.018688267096877098,
0.0504012256860733,
0.050729844719171524,
0.03199297562241554,
-0.05676629766821861,
0.04105515405535698,
0.011234153993427753,
-0.029026774689555168,
-0.1420264095067978,
-0.0753667801618576,
0.055518005043268204,
-0.001751455245539546,
0.043009642511606216,
-0.05537867173552513,
0.027575301006436348,
0.13269533216953278,
-0.008434117771685123,
-0.06712769716978073,
0.07141239941120148,
-0.07733511924743652,
0.09029383957386017,
-0.013053758069872856,
-0.03516661003232002,
-0.050079189240932465,
-0.006445335689932108,
0.03310399875044823,
0.011884858831763268,
-0.033930227160453796,
-0.019224366173148155,
-0.04618232324719429,
-0.04731970652937889,
0.20120128989219666,
-0.021115614101290703,
-0.15876077115535736,
0.002861914923414588,
-0.09207534044981003,
0.12609347701072693,
-0.018734483048319817,
0.14374245703220367,
0.19815626740455627,
0.024382619187235832,
0.008123079314827919,
-0.04330689087510109,
0.034939441829919815,
0.007495005615055561,
-0.04539262503385544,
-0.0846559926867485
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-large-cased-bn-adapter-3.17M-snli-model1
This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7500
- Accuracy: 0.736
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 83
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.412 | 1.0 | 8584 | 0.3365 | 0.8770 |
| 0.3706 | 2.0 | 17168 | 0.3072 | 0.8872 |
| 0.3597 | 3.0 | 25752 | 0.2988 | 0.8888 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-large-cased", "model-index": [{"name": "bert-large-cased-bn-adapter-3.17M-snli-model1", "results": []}]} | null | varun-v-rao/bert-large-cased-bn-adapter-3.17M-snli-model1 | [
"tensorboard",
"generated_from_trainer",
"base_model:bert-large-cased",
"license:apache-2.0",
"region:us"
] | 2024-02-06T04:34:08+00:00 | [] | [] | TAGS
#tensorboard #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #region-us
| bert-large-cased-bn-adapter-3.17M-snli-model1
=============================================
This model is a fine-tuned version of bert-large-cased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7500
* Accuracy: 0.736
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 83
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 83\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#tensorboard #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 83\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
37,
98,
4,
33
] | [
"passage: TAGS\n#tensorboard #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 83\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.09240599721670151,
0.008305436931550503,
-0.0009757168008945882,
0.10874024778604507,
0.1635274738073349,
0.03409658372402191,
0.14058178663253784,
0.0902060866355896,
-0.12397259473800659,
0.019144149497151375,
0.10259444266557693,
0.10818060487508774,
0.009139712899923325,
0.07315446436405182,
-0.022586921229958534,
-0.2367015928030014,
-0.011283166706562042,
0.03786800801753998,
-0.12947320938110352,
0.11912231892347336,
0.0885123759508133,
-0.1360803246498108,
0.07304488867521286,
0.0074118925258517265,
-0.25938722491264343,
0.04532434791326523,
0.01284931693226099,
-0.05351780727505684,
0.13299542665481567,
-0.0026367197278887033,
0.15917779505252838,
-0.011397544294595718,
0.11992819607257843,
-0.15321868658065796,
0.01121707446873188,
0.07825954258441925,
0.00958416797220707,
0.08225493878126144,
0.05342646688222885,
0.030429713428020477,
0.11104167997837067,
-0.08401809632778168,
0.055811602622270584,
0.028017321601510048,
-0.12472350895404816,
-0.2690873146057129,
-0.0833154022693634,
0.01059191394597292,
0.05958983674645424,
0.08579158037900925,
0.00017526163719594479,
0.14592304825782776,
-0.08744151890277863,
0.07767187803983688,
0.2677844166755676,
-0.264871746301651,
-0.0789894312620163,
0.08610714972019196,
0.01306493952870369,
0.09809256345033646,
-0.10175397992134094,
-0.03346711024641991,
0.08245495706796646,
0.04586203768849373,
0.11260928213596344,
-0.026858285069465637,
-0.1438034176826477,
0.04389190301299095,
-0.150065079331398,
-0.0021878264378756285,
0.09918612241744995,
0.03699416667222977,
-0.026401611045002937,
0.016609415411949158,
-0.07899125665426254,
-0.13581418991088867,
-0.04605001211166382,
-0.020138556137681007,
0.07903831452131271,
-0.035009317100048065,
-0.09631150215864182,
-0.031099751591682434,
-0.1071450263261795,
-0.10969632118940353,
-0.0580771341919899,
0.16630792617797852,
0.04453010484576225,
0.048851724714040756,
-0.0655675157904625,
0.11337506026029587,
-0.06276587396860123,
-0.136363685131073,
0.038381364196538925,
0.037876859307289124,
-0.006529466714709997,
-0.047216080129146576,
-0.07133421301841736,
-0.07104356586933136,
0.028543313965201378,
0.1267615109682083,
-0.08412796258926392,
0.0473647266626358,
0.035128843039274216,
0.04851942136883736,
-0.1102699488401413,
0.18097329139709473,
-0.05922798812389374,
0.025426818057894707,
0.00024229407426901162,
0.06402073800563812,
0.011845838278532028,
-0.005357719957828522,
-0.11116786301136017,
0.025055186823010445,
0.07104349881410599,
0.009301707148551941,
-0.08702002465724945,
0.023243600502610207,
-0.032886963337659836,
-0.00695532513782382,
-0.006663133390247822,
-0.09356138110160828,
0.04113706946372986,
0.0029370326083153486,
-0.09537278860807419,
-0.012128487229347229,
0.009807904250919819,
0.02748226933181286,
0.018018411472439766,
0.11478358507156372,
-0.09649809449911118,
0.05789397284388542,
-0.13548126816749573,
-0.10460185259580612,
0.012393434531986713,
-0.04672931134700775,
0.021675340831279755,
-0.06395318359136581,
-0.1651175618171692,
-0.0013707459438592196,
0.08024939894676208,
-0.04278113692998886,
-0.02413511462509632,
-0.02663019672036171,
-0.09343581646680832,
-0.004095264710485935,
-0.015757646411657333,
0.19884899258613586,
-0.05513778701424599,
0.09330077469348907,
0.0645723044872284,
0.06510944664478302,
-0.0961228609085083,
0.036491088569164276,
-0.08887647092342377,
0.008503327146172523,
-0.24113944172859192,
0.010633048601448536,
-0.06624115258455276,
0.07215952128171921,
-0.07971572130918503,
-0.08363930135965347,
0.007347145117819309,
0.0027129575610160828,
0.1026940643787384,
0.05975985899567604,
-0.21051029860973358,
-0.07098828256130219,
0.1277288943529129,
-0.07740181684494019,
-0.10733508318662643,
0.09506798535585403,
-0.053590863943099976,
0.0690034031867981,
0.06329724192619324,
0.17195415496826172,
-0.023958489298820496,
-0.1061847060918808,
0.025066543370485306,
-0.015998506918549538,
0.07227089256048203,
-0.09272593259811401,
0.059665925800800323,
-0.02807946316897869,
0.01049904152750969,
0.027694009244441986,
-0.05402632802724838,
0.04954315721988678,
-0.105466328561306,
-0.09330065548419952,
-0.06608543545007706,
-0.11010228097438812,
-0.006105748470872641,
0.0705442801117897,
0.08169668912887573,
-0.10502259433269501,
-0.07708263397216797,
0.1299496442079544,
0.08703234791755676,
-0.05155874043703079,
0.03230710327625275,
-0.04881122335791588,
0.058817677199840546,
-0.07153147459030151,
-0.031560271978378296,
-0.18343795835971832,
-0.026107164099812508,
0.012308932840824127,
-0.02364921011030674,
0.050820644944906235,
-0.0010627915617078543,
0.08506029844284058,
0.06457850337028503,
-0.06192568689584732,
0.008464218117296696,
-0.07800345867872238,
-0.01065357681363821,
-0.1317218840122223,
-0.21550299227237701,
-0.038341086357831955,
-0.00929124467074871,
0.06849588453769684,
-0.19433295726776123,
0.026972822844982147,
-0.021354004740715027,
0.08623296022415161,
0.005167899653315544,
-0.007354323752224445,
-0.06892212480306625,
0.0883849486708641,
-0.0011190089862793684,
-0.06984598934650421,
0.04520973563194275,
-0.013898611068725586,
-0.0683196559548378,
-0.08442830294370651,
-0.07577432692050934,
0.15233688056468964,
0.143630251288414,
-0.11697940528392792,
-0.06233878433704376,
0.03442743048071861,
-0.06773699074983597,
-0.034421760588884354,
-0.05486934632062912,
0.05324002355337143,
0.1724032759666443,
-0.011278605088591576,
0.12879303097724915,
-0.09767697006464005,
-0.044616151601076126,
0.005193696822971106,
-0.024743512272834778,
0.06142768636345863,
0.08256204426288605,
0.17339998483657837,
-0.03936966881155968,
0.11356319487094879,
0.16854040324687958,
-0.1534477025270462,
0.11573775112628937,
-0.07803870737552643,
-0.09580191224813461,
-0.023065945133566856,
-0.028943587094545364,
-0.008024499751627445,
0.16945219039916992,
-0.0776345506310463,
0.007910252548754215,
0.0025800741277635098,
0.018120717257261276,
0.049364954233169556,
-0.24715028703212738,
-0.03721645846962929,
0.013023439794778824,
-0.05045006051659584,
-0.023122277110815048,
-0.03216147795319557,
0.006979696918278933,
0.09382878988981247,
-0.032793983817100525,
-0.07276993989944458,
0.019907353445887566,
0.005397424101829529,
-0.055344514548778534,
0.20955997705459595,
-0.07164476066827774,
-0.0428021065890789,
-0.11590204387903214,
-0.03279783949255943,
-0.053609367460012436,
-0.022693628445267677,
0.04169217124581337,
-0.13050146400928497,
-0.031903382390737534,
-0.05523088201880455,
0.03478191792964935,
0.022741517052054405,
0.04601479694247246,
0.023648494854569435,
0.007654099725186825,
0.09840958565473557,
-0.14360438287258148,
0.021882249042391777,
-0.07209320366382599,
-0.09564709663391113,
0.011628593318164349,
0.07657929509878159,
0.11159619688987732,
0.15920211374759674,
-0.023250866681337357,
0.005550620146095753,
-0.009231828153133392,
0.256466269493103,
-0.06089124083518982,
-0.0389949232339859,
0.11359646916389465,
-0.0020373344887048006,
0.051398683339357376,
0.09578216075897217,
0.10470783710479736,
-0.11064618080854416,
-0.0036092561203986406,
0.05874362960457802,
-0.0388830229640007,
-0.24244490265846252,
-0.005823533516377211,
-0.030735347419977188,
-0.06716552376747131,
0.0727798342704773,
0.03920756280422211,
0.04370053857564926,
0.06037183105945587,
0.056046582758426666,
0.10411234945058823,
-0.06922370195388794,
0.0415794812142849,
0.06053385138511658,
0.06628987938165665,
0.1094207689166069,
-0.061187196522951126,
-0.05849846452474594,
0.03453635796904564,
-0.003722403896972537,
0.24994036555290222,
0.005089092534035444,
0.13869808614253998,
0.07061063498258591,
0.18170104920864105,
-0.013406896032392979,
0.0975048691034317,
-0.027569480240345,
-0.05944187939167023,
-0.0035323710180819035,
-0.059538714587688446,
0.022459544241428375,
0.019916007295250893,
-0.09024001657962799,
0.07523693889379501,
-0.075617216527462,
-0.0004019620828330517,
0.06006208062171936,
0.25965678691864014,
0.0019539890345185995,
-0.2972118556499481,
-0.03862421587109566,
0.0002637439174577594,
-0.009067357517778873,
-0.006948401220142841,
0.01583930291235447,
0.12365422397851944,
-0.032839540392160416,
0.009863284416496754,
-0.048863060772418976,
0.09760219603776932,
0.011361462995409966,
0.028651360422372818,
0.07950767874717712,
0.16561266779899597,
-0.0013570637675002217,
0.05889934301376343,
-0.29690656065940857,
0.28026217222213745,
0.030196228995919228,
0.11595796048641205,
-0.054777972400188446,
-0.0019855601713061333,
0.03552338480949402,
0.02042223885655403,
0.015188266523182392,
-0.024546924978494644,
-0.04895959049463272,
-0.16517029702663422,
-0.06301689147949219,
0.05472341552376747,
0.10786398500204086,
-0.0043727015145123005,
0.08134961873292923,
-0.006586093455553055,
0.014885934069752693,
0.10187920928001404,
-0.04906077682971954,
-0.1288476288318634,
-0.05525169149041176,
-0.056734319776296616,
0.027199523523449898,
-0.10319698601961136,
-0.08283726871013641,
-0.11198693513870239,
-0.09762099385261536,
0.10040551424026489,
0.023126984015107155,
-0.02188815362751484,
-0.10852834582328796,
0.09509537369012833,
0.1245664581656456,
-0.05008234828710556,
0.04192127287387848,
0.022971762344241142,
0.05357440188527107,
0.03502862900495529,
-0.06150687858462334,
0.11518438905477524,
-0.055274128913879395,
-0.15001735091209412,
-0.04823426157236099,
0.06728770583868027,
0.0480022206902504,
0.054347191005945206,
-0.03498940169811249,
0.012115687131881714,
-0.017974944785237312,
-0.10353229939937592,
0.010129019618034363,
-0.048176076263189316,
0.031231794506311417,
0.0010164605919271708,
-0.03191184997558594,
0.034474942833185196,
-0.04576345160603523,
-0.02699725143611431,
0.12609811127185822,
0.27415207028388977,
-0.07741783559322357,
-0.015360422432422638,
0.06319595873355865,
-0.06449741125106812,
-0.15793359279632568,
0.08787965029478073,
0.0663323774933815,
-0.002351314527913928,
0.04553171247243881,
-0.17983394861221313,
0.17198051512241364,
0.13163577020168304,
-0.01726481504738331,
0.12746849656105042,
-0.3216763734817505,
-0.14790402352809906,
0.10498281568288803,
0.17208077013492584,
0.1146472841501236,
-0.15626534819602966,
-0.02897326461970806,
-0.02764793671667576,
-0.11806687712669373,
0.10506681352853775,
-0.17840275168418884,
0.09593887627124786,
-0.014462956227362156,
0.09366042166948318,
-0.018440527841448784,
-0.06070968508720398,
0.14248433709144592,
0.019289251416921616,
0.13236501812934875,
-0.03180575370788574,
-0.003144502406939864,
0.10617683827877045,
-0.02651205286383629,
0.010456814430654049,
-0.06551176309585571,
0.02178196609020233,
-0.03750763088464737,
-0.0035651354119181633,
-0.0738861933350563,
0.03490782529115677,
-0.0438416562974453,
-0.06138481944799423,
-0.0482555590569973,
0.008011456578969955,
-0.007169393356889486,
-0.02658778429031372,
0.13018257915973663,
0.05899728462100029,
0.14232923090457916,
0.10824451595544815,
0.033293116837739944,
-0.08706790208816528,
-0.06354095786809921,
0.023822108283638954,
-0.014658666215837002,
0.0834975317120552,
-0.19075722992420197,
0.004581124056130648,
0.11962885409593582,
0.027370184659957886,
0.09400611370801926,
0.08897623419761658,
-0.06292764097452164,
0.012369035743176937,
0.07559016346931458,
-0.1383303850889206,
-0.08538196235895157,
0.016587311401963234,
-0.041864603757858276,
-0.10181734710931778,
0.06409421563148499,
0.06555549800395966,
-0.07381585985422134,
-0.006480835843831301,
-0.002388437045738101,
-0.020565753802657127,
-0.09491118043661118,
0.21391059458255768,
0.09318733215332031,
0.04764675721526146,
-0.11343421041965485,
0.09785373508930206,
0.05214862897992134,
-0.03712471202015877,
-0.0037447665818035603,
0.07294440269470215,
-0.04771391674876213,
0.005924853961914778,
0.14498791098594666,
0.22478587925434113,
-0.06769604235887527,
-0.06369268149137497,
-0.1686529666185379,
-0.09489999711513519,
0.052817992866039276,
0.1814207285642624,
0.11325260996818542,
-0.012439591810107231,
-0.03250417113304138,
0.045354463160037994,
-0.12976153194904327,
0.06912378966808319,
0.01744901016354561,
0.08496064692735672,
-0.13167446851730347,
0.13042216002941132,
0.037615109235048294,
0.01747467741370201,
-0.01864521950483322,
0.05439029261469841,
-0.11671888083219528,
0.024172095581889153,
-0.14870524406433105,
-0.052832309156656265,
0.021127616986632347,
-0.004606792237609625,
0.012340952642261982,
-0.07347185164690018,
-0.09033741801977158,
0.03544159606099129,
-0.1372196227312088,
-0.021390685811638832,
0.05346791446208954,
0.03733094409108162,
-0.14561209082603455,
-0.012417883612215519,
0.02599327638745308,
-0.054441362619400024,
0.019187303259968758,
0.040186360478401184,
0.032506994903087616,
0.09315267950296402,
-0.17142057418823242,
-0.011211831122636795,
0.05504956096410751,
0.009329847991466522,
0.09107495099306107,
-0.03820420429110527,
-0.030444929376244545,
-0.011356758885085583,
0.1275160163640976,
0.000297351652989164,
0.056150320917367935,
-0.13243122398853302,
-0.016802018508315086,
-0.05252525582909584,
-0.08659142255783081,
-0.035562239587306976,
0.0038103775586932898,
0.10365480929613113,
0.03171781450510025,
0.18199360370635986,
-0.06156018003821373,
0.010475310496985912,
-0.22957107424736023,
0.001876913825981319,
-0.016263607889413834,
-0.07712298631668091,
-0.11100330203771591,
-0.0473109669983387,
0.055086929351091385,
-0.07011011987924576,
0.1592523455619812,
0.018734388053417206,
0.044785454869270325,
0.03620930761098862,
-0.010028550401329994,
0.004783002194017172,
0.03440574184060097,
0.252330482006073,
0.04416511580348015,
0.00015889090718701482,
0.06908519566059113,
0.06266501545906067,
0.115864597260952,
0.08933970332145691,
0.2088521271944046,
0.15028877556324005,
-0.025138359516859055,
0.11413311213254929,
0.06935875862836838,
-0.0677843764424324,
-0.10431282222270966,
0.0905727744102478,
-0.029912235215306282,
0.06662863492965698,
-0.025072090327739716,
0.22129444777965546,
0.09155666083097458,
-0.16235767304897308,
0.03605179488658905,
-0.04000413790345192,
-0.0902760922908783,
-0.10888074338436127,
0.02419796958565712,
-0.07499681413173676,
-0.1858646124601364,
0.02158709242939949,
-0.10684576630592346,
0.011643794365227222,
0.15803837776184082,
0.001822295249439776,
0.007549960166215897,
0.19001352787017822,
0.05845081806182861,
0.05919674038887024,
0.025849008932709694,
-0.0038764497730880976,
-0.04435371235013008,
-0.08816713839769363,
-0.07420595735311508,
-0.012006058357656002,
-0.016819428652524948,
0.031003426760435104,
-0.06318092346191406,
-0.08408389240503311,
0.0465775802731514,
0.003117719665169716,
-0.0870688408613205,
0.019590258598327637,
0.011985380202531815,
0.06556180864572525,
0.017905544489622116,
0.006332474760711193,
0.019406858831644058,
-0.02045077085494995,
0.2056000530719757,
-0.08455845713615417,
-0.10850859433412552,
-0.08930835872888565,
0.2103743553161621,
0.011993746273219585,
-0.009025968611240387,
0.013977941125631332,
-0.08928091824054718,
-0.017485855147242546,
0.20538382232189178,
0.17087417840957642,
-0.10897123068571091,
-0.03374992683529854,
-0.008316517807543278,
-0.02350146695971489,
-0.09482266753911972,
0.12391338497400284,
0.13058564066886902,
0.021356968209147453,
-0.11565475910902023,
-0.05442352965474129,
-0.054555218666791916,
-0.022342059761285782,
-0.051518362015485764,
0.03980175778269768,
0.04866849258542061,
-0.0013014792930334806,
-0.04748525843024254,
0.0630558580160141,
-0.047054581344127655,
-0.10256846994161606,
0.0743049681186676,
-0.15957333147525787,
-0.17613576352596283,
-0.010963126085698605,
0.1525265872478485,
-0.025585750117897987,
0.0553346648812294,
-0.043057311326265335,
0.0023505345452576876,
0.0471741259098053,
-0.04239291325211525,
-0.093897745013237,
-0.13210760056972504,
0.11153151094913483,
-0.13416996598243713,
0.22810767590999603,
-0.03508787974715233,
0.10399531573057175,
0.11457673460245132,
0.044015899300575256,
-0.07592125236988068,
0.047820787876844406,
0.036477167159318924,
-0.12807199358940125,
-0.016626177355647087,
0.0961654931306839,
-0.039514731615781784,
0.04592396691441536,
0.021922126412391663,
-0.11442621052265167,
0.013687650673091412,
-0.02781451679766178,
-0.04481799155473709,
-0.05489412322640419,
-0.08512888103723526,
-0.08106500655412674,
0.09921231865882874,
0.19218608736991882,
-0.020752090960741043,
0.03455611318349838,
-0.08417476713657379,
0.048724450170993805,
0.07060018926858902,
0.03792481869459152,
-0.07315272837877274,
-0.23552361130714417,
0.027973558753728867,
0.10198929905891418,
-0.06113945692777634,
-0.2176131159067154,
-0.08226500451564789,
0.03227808326482773,
-0.07863735407590866,
-0.08446099609136581,
0.06759468466043472,
0.12880966067314148,
0.07018348574638367,
-0.05074905604124069,
-0.12791596353054047,
-0.09773427248001099,
0.13571231067180634,
-0.1607576608657837,
-0.09971068054437637
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-bn-adapter-895K-snli-model1
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7623
- Accuracy: 0.728
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 61
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.4254 | 1.0 | 8584 | 0.3365 | 0.8722 |
| 0.4021 | 2.0 | 17168 | 0.3165 | 0.8790 |
| 0.3806 | 3.0 | 25752 | 0.3115 | 0.8817 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-bn-adapter-895K-snli-model1", "results": []}]} | null | varun-v-rao/roberta-base-bn-adapter-895K-snli-model1 | [
"tensorboard",
"generated_from_trainer",
"base_model:roberta-base",
"license:mit",
"region:us"
] | 2024-02-06T04:35:02+00:00 | [] | [] | TAGS
#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us
| roberta-base-bn-adapter-895K-snli-model1
========================================
This model is a fine-tuned version of roberta-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7623
* Accuracy: 0.728
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 61
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 61\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 61\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
31,
98,
4,
33
] | [
"passage: TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-base #license-mit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 61\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.07504717260599136,
-0.02196611650288105,
-0.0010848628589883447,
0.11662556231021881,
0.19939157366752625,
0.03200864791870117,
0.1486678421497345,
0.06581471115350723,
-0.129006028175354,
0.013607422821223736,
0.10531160980463028,
0.1320256143808365,
-0.01142910122871399,
0.055550456047058105,
-0.021502450108528137,
-0.26036733388900757,
-0.030642813071608543,
0.03408229723572731,
-0.14363013207912445,
0.12668326497077942,
0.08404766023159027,
-0.14420217275619507,
0.07034140825271606,
-0.0012516307178884745,
-0.3223230242729187,
0.053184445947408676,
0.0346849150955677,
-0.041734255850315094,
0.14334385097026825,
-0.008378831669688225,
0.1929733157157898,
-0.01450236327946186,
0.13221895694732666,
-0.1275121122598648,
0.013382394798099995,
0.06838522851467133,
0.0015898996498435736,
0.08297046273946762,
0.05717143416404724,
0.02530883252620697,
0.11438415944576263,
-0.0843142420053482,
0.06268257647752762,
0.019262339919805527,
-0.14505407214164734,
-0.24430865049362183,
-0.06961145997047424,
-0.027241315692663193,
0.054433126002550125,
0.08689621835947037,
-0.010947417467832565,
0.15181951224803925,
-0.09311173111200333,
0.09624715894460678,
0.2703189253807068,
-0.24810239672660828,
-0.0832735151052475,
0.08836237341165543,
0.003767068963497877,
0.09513412415981293,
-0.10942121595144272,
-0.027416694909334183,
0.07367566227912903,
0.04249269515275955,
0.12078028917312622,
-0.03255759924650192,
-0.1243315190076828,
0.04483936354517937,
-0.15376335382461548,
0.010548665188252926,
0.06227191165089607,
0.02109299600124359,
-0.01804467663168907,
0.047883693128824234,
-0.08472201228141785,
-0.1384362280368805,
-0.05161247402429581,
-0.021206233650445938,
0.08348450064659119,
-0.051465120166540146,
-0.12976142764091492,
-0.020225699990987778,
-0.10497960448265076,
-0.10317029058933258,
-0.06085487827658653,
0.2162429690361023,
0.0503382608294487,
0.040684036910533905,
-0.06791013479232788,
0.11257990449666977,
-0.059102773666381836,
-0.13090306520462036,
0.05595609173178673,
0.05160818248987198,
-0.024088367819786072,
-0.058891989290714264,
-0.08170733600854874,
-0.09781599044799805,
0.017064997926354408,
0.11480006575584412,
-0.09262832254171371,
0.051889292895793915,
0.05852816253900528,
0.0545879490673542,
-0.10206030309200287,
0.1756538599729538,
-0.04924171790480614,
0.03398660570383072,
0.007845277898013592,
0.02624281495809555,
-0.005575884599238634,
0.0010153430048376322,
-0.113545261323452,
0.022890591993927956,
0.07911255210638046,
-0.0018595104338601232,
-0.08761811256408691,
0.036595605313777924,
-0.028723018243908882,
-0.002428269712254405,
-0.03249514847993851,
-0.08729509264230728,
0.047695092856884,
-0.004337730351835489,
-0.09814488142728806,
-0.019794931635260582,
0.009031648747622967,
0.015232909470796585,
0.026923397555947304,
0.13925957679748535,
-0.11064159125089645,
0.08269103616476059,
-0.14556868374347687,
-0.12698261439800262,
-0.004967650398612022,
-0.04249296709895134,
0.027514895424246788,
-0.07261972874403,
-0.14282295107841492,
-0.015082732774317265,
0.06737703084945679,
-0.03969414904713631,
-0.008101433515548706,
-0.03485671430826187,
-0.09631557762622833,
-0.014287775382399559,
-0.009238864295184612,
0.1879095733165741,
-0.045073509216308594,
0.10581435263156891,
0.060734823346138,
0.06837314367294312,
-0.10672135651111603,
0.04194086417555809,
-0.0915866270661354,
0.0037800343707203865,
-0.2695055305957794,
0.011428271420300007,
-0.05232617259025574,
0.0656001940369606,
-0.06637563556432724,
-0.1009121984243393,
0.020047051832079887,
0.001043503638356924,
0.110341377556324,
0.07093962281942368,
-0.19003789126873016,
-0.07768629491329193,
0.1225532814860344,
-0.07452914863824844,
-0.06938943266868591,
0.08727103471755981,
-0.07499042898416519,
0.07354842871427536,
0.06236662715673447,
0.19436030089855194,
-0.04038422927260399,
-0.10317637026309967,
0.04535488784313202,
-0.03786056116223335,
0.03933946415781975,
-0.08405590802431107,
0.03961080685257912,
-0.009299595840275288,
0.011383473873138428,
0.02626683935523033,
-0.05900359898805618,
0.06742952764034271,
-0.13039334118366241,
-0.07820023596286774,
-0.04925829544663429,
-0.09391389787197113,
0.021193066611886024,
0.07966858148574829,
0.1003238782286644,
-0.10288625955581665,
-0.055121030658483505,
0.18817169964313507,
0.05893944576382637,
-0.05473455786705017,
0.03007838875055313,
-0.034951578825712204,
0.04435337707400322,
-0.07125015556812286,
-0.045315518975257874,
-0.20013868808746338,
-0.051579009741544724,
0.003357405075803399,
0.004871206823736429,
0.07278043031692505,
0.022595029324293137,
0.09174831956624985,
0.07610823214054108,
-0.07088521867990494,
0.015233958140015602,
-0.08101925998926163,
-0.0032595437951385975,
-0.13743820786476135,
-0.22669917345046997,
-0.03477611020207405,
-0.011184073984622955,
0.07211963832378387,
-0.19533808529376984,
0.0282870102673769,
-0.0375007688999176,
0.07303834706544876,
0.0017272054683417082,
-0.023834966123104095,
-0.0730145126581192,
0.10271139442920685,
0.004017908591777086,
-0.06553107500076294,
0.048573095351457596,
-0.016662921756505966,
-0.06672792136669159,
-0.09143415838479996,
-0.08729708194732666,
0.1717613786458969,
0.13816849887371063,
-0.16804343461990356,
-0.10477761179208755,
0.032801538705825806,
-0.06593187153339386,
-0.026494517922401428,
-0.06033870577812195,
0.0610102117061615,
0.18518884479999542,
-0.016422197222709656,
0.13672617077827454,
-0.08015266805887222,
-0.03968430310487747,
0.0014118484687060118,
-0.04536481574177742,
0.07874511182308197,
0.09738478064537048,
0.18343159556388855,
-0.0590268149971962,
0.11254315078258514,
0.14170686900615692,
-0.15538613498210907,
0.1261107623577118,
-0.05953202396631241,
-0.07385005056858063,
-0.04317384958267212,
-0.026662718504667282,
-0.009321928024291992,
0.15686804056167603,
-0.06881535053253174,
0.006760594435036182,
-0.012581145390868187,
0.012095834128558636,
0.04155924543738365,
-0.2530233561992645,
-0.05647828057408333,
0.013558820821344852,
-0.029919032007455826,
-0.019113775342702866,
-0.03796585276722908,
0.008048737421631813,
0.10730927437543869,
-0.03022783063352108,
-0.07068441063165665,
0.014685403555631638,
0.01607350818812847,
-0.05794588848948479,
0.22635643184185028,
-0.062434837222099304,
-0.0056570107117295265,
-0.09861915558576584,
-0.035421162843704224,
-0.04823165386915207,
-0.010285736992955208,
0.03573666885495186,
-0.1389041692018509,
-0.018096521496772766,
-0.03202005848288536,
0.0484643317759037,
0.0001384701899951324,
0.047474708408117294,
-0.014051380567252636,
0.006750885397195816,
0.07721604406833649,
-0.13748839497566223,
0.017168357968330383,
-0.09384776651859283,
-0.10211171209812164,
0.03919946402311325,
0.07640279829502106,
0.12626564502716064,
0.169523686170578,
-0.0357414148747921,
0.0015493450919166207,
-0.011123191565275192,
0.2794734537601471,
-0.0656852126121521,
-0.04077553376555443,
0.09294551610946655,
0.001982476096600294,
0.05177205428481102,
0.09133914113044739,
0.11456448584794998,
-0.11424130201339722,
-0.004628014750778675,
0.05282793194055557,
-0.05832276865839958,
-0.22508031129837036,
-0.012869362719357014,
-0.024534175172448158,
-0.10043153166770935,
0.06033077836036682,
0.029125049710273743,
0.023160334676504135,
0.05973955988883972,
0.07412649691104889,
0.11056091636419296,
-0.0724787563085556,
0.0450228787958622,
0.05505984276533127,
0.06084202975034714,
0.11306687444448471,
-0.05008183792233467,
-0.08496854454278946,
0.01885630376636982,
-0.039204128086566925,
0.2729382812976837,
0.007082182914018631,
0.11201981455087662,
0.07496020197868347,
0.17599065601825714,
-0.004768769256770611,
0.09151674807071686,
0.0043555148877203465,
-0.06639955937862396,
-0.005801725201308727,
-0.050829045474529266,
0.021919455379247665,
0.012182170525193214,
-0.08405732363462448,
0.05270947888493538,
-0.07741663604974747,
-0.018180305138230324,
0.05304606631398201,
0.21042151749134064,
-0.011892623268067837,
-0.3013734221458435,
-0.04410148784518242,
-0.016421958804130554,
0.004198143258690834,
0.004865860100835562,
-0.0021876811515539885,
0.13577327132225037,
-0.04655320197343826,
-0.004250322934240103,
-0.05897257849574089,
0.08524153381586075,
-0.0051977066323161125,
0.03135285899043083,
0.06274611502885818,
0.16399435698986053,
-0.018998289480805397,
0.044199515134096146,
-0.2988342344760895,
0.30715441703796387,
0.02464265748858452,
0.10885931551456451,
-0.037599071860313416,
-0.019467920064926147,
0.016843190416693687,
0.028837164863944054,
0.002446346450597048,
-0.018593449145555496,
-0.051433056592941284,
-0.21441519260406494,
-0.028060778975486755,
0.05275378376245499,
0.13296277821063995,
0.010890397243201733,
0.08687815070152283,
0.005421625450253487,
0.014498201198875904,
0.10397488623857498,
-0.05973513796925545,
-0.1091231182217598,
-0.030007638037204742,
-0.0649598017334938,
0.001186413923278451,
-0.10763458907604218,
-0.05461598560214043,
-0.10675837099552155,
-0.0876021608710289,
0.09862033277750015,
0.022322356700897217,
-0.029116380959749222,
-0.12454208731651306,
0.1378069818019867,
0.10862619429826736,
-0.06231934204697609,
0.04682883247733116,
0.024252798408269882,
0.03225237876176834,
0.03843381628394127,
-0.05745267868041992,
0.12105092406272888,
-0.042894311249256134,
-0.13914915919303894,
-0.051618967205286026,
0.0450114831328392,
0.05964935943484306,
0.0641997829079628,
-0.03370054066181183,
0.023258181288838387,
-0.017583021894097328,
-0.11111718416213989,
0.026309940963983536,
-0.06602529436349869,
0.02015288732945919,
0.008357956074178219,
-0.03415895253419876,
0.05641811713576317,
-0.042795781046152115,
-0.0017601674189791083,
0.1227617859840393,
0.2873076796531677,
-0.07966960221529007,
-0.03170030191540718,
0.04599546641111374,
-0.07222023606300354,
-0.16022683680057526,
0.11501919478178024,
0.08033297955989838,
-0.01233722735196352,
0.07074001431465149,
-0.17645403742790222,
0.15671542286872864,
0.1304551064968109,
-0.01038731075823307,
0.11059895157814026,
-0.3227219879627228,
-0.15281713008880615,
0.07695726305246353,
0.20852673053741455,
0.16068153083324432,
-0.1567164808511734,
-0.017311174422502518,
-0.021646948531270027,
-0.09880055487155914,
0.09754623472690582,
-0.15074068307876587,
0.10705668479204178,
-0.011216029524803162,
0.1066756621003151,
-0.007621145807206631,
-0.06650811433792114,
0.11953500658273697,
0.0271914005279541,
0.1516515463590622,
-0.04045558720827103,
-0.04140360653400421,
0.10805528610944748,
-0.009583677165210247,
-0.009147804230451584,
-0.012648473493754864,
0.011384568177163601,
-0.02640996128320694,
0.0008434430928900838,
-0.08612505346536636,
0.044789064675569534,
-0.03805799037218094,
-0.07103253901004791,
-0.051249947398900986,
-0.00032585079316049814,
-0.012952707707881927,
-0.027851371094584465,
0.12841936945915222,
0.05308148264884949,
0.15231244266033173,
0.08748439699411392,
0.0027266456745564938,
-0.050451699644327164,
-0.0697089284658432,
0.024458637461066246,
-0.0022189421579241753,
0.07547595351934433,
-0.18637779355049133,
-0.011120353825390339,
0.12509438395500183,
0.0432291105389595,
0.09337848424911499,
0.10056798905134201,
-0.05434801056981087,
0.0336456373333931,
0.06764329224824905,
-0.14672289788722992,
-0.09095470607280731,
0.02378864400088787,
-0.1047607958316803,
-0.06498025357723236,
0.061855118721723557,
0.07627268135547638,
-0.08077145367860794,
-0.0029192804358899593,
-0.01171997468918562,
-0.03554819896817207,
-0.08522024005651474,
0.22624783217906952,
0.11208636313676834,
0.037214115262031555,
-0.11572107672691345,
0.07430356740951538,
0.056528568267822266,
-0.018854454159736633,
-0.00003477922291494906,
0.07410172373056412,
-0.04799585044384003,
0.012807918712496758,
0.15142543613910675,
0.24081581830978394,
-0.11076168715953827,
-0.02714712545275688,
-0.19115282595157623,
-0.08052684366703033,
0.041816096752882004,
0.19154661893844604,
0.1151590347290039,
-0.02147534303367138,
-0.028267724439501762,
0.047060199081897736,
-0.15340858697891235,
0.05188770592212677,
0.011731041595339775,
0.09047716856002808,
-0.1140303835272789,
0.18979203701019287,
0.03028956428170204,
0.01859944500029087,
-0.0310561154037714,
0.0498436763882637,
-0.1481277346611023,
0.030484048649668694,
-0.10556831955909729,
-0.06954585760831833,
0.028942665085196495,
-0.018184753134846687,
0.0040140897035598755,
-0.0799974650144577,
-0.07830949127674103,
0.034562867134809494,
-0.14538533985614777,
-0.014190074987709522,
0.05907344073057175,
0.012697918340563774,
-0.14707835018634796,
-0.01859811320900917,
0.025801798328757286,
-0.04319806769490242,
0.02280118502676487,
0.05511784553527832,
0.0365181490778923,
0.0937771201133728,
-0.1962488889694214,
-0.010923209600150585,
0.06608492881059647,
-0.013938053511083126,
0.09249706566333771,
-0.021204089745879173,
-0.027661392465233803,
-0.028898559510707855,
0.15228873491287231,
0.009173423051834106,
0.06013000011444092,
-0.13488399982452393,
0.00866401195526123,
-0.05941062793135643,
-0.09692855179309845,
-0.043903809040784836,
0.0030080038122832775,
0.08204370737075806,
0.0315031073987484,
0.15750204026699066,
-0.07005921751260757,
0.027573361992836,
-0.22877438366413116,
-0.016743099316954613,
-0.0053360979072749615,
-0.08171126991510391,
-0.10456497222185135,
-0.049788735806941986,
0.0663323625922203,
-0.05929097160696983,
0.13708552718162537,
0.05557160824537277,
0.061962343752384186,
0.03766225650906563,
0.020741809159517288,
0.00816123653203249,
0.029697027057409286,
0.22349712252616882,
0.038620930165052414,
-0.01827237941324711,
0.06854332238435745,
0.09894901514053345,
0.11607451736927032,
0.08517762273550034,
0.21238110959529877,
0.14627605676651,
-0.058509424328804016,
0.09536143392324448,
0.06390983611345291,
-0.04759204760193825,
-0.11115319281816483,
0.08345220237970352,
-0.021572943776845932,
0.06151537224650383,
-0.03247610107064247,
0.20591102540493011,
0.07779636979103088,
-0.15807965397834778,
0.03848204016685486,
-0.03396168351173401,
-0.10101024061441422,
-0.10189927369356155,
0.042317334562540054,
-0.07779209315776825,
-0.17601418495178223,
0.03504127636551857,
-0.10349906980991364,
0.019131124019622803,
0.1568201631307602,
-0.0031464998610317707,
0.008439913392066956,
0.19781845808029175,
0.08514557778835297,
0.053798872977495193,
0.017570916563272476,
-0.0060579776763916016,
-0.0412072129547596,
-0.0898013636469841,
-0.06987423449754715,
-0.020744314417243004,
-0.02024741843342781,
0.01685757376253605,
-0.05432429909706116,
-0.0972810834646225,
0.036543797701597214,
-0.0032237630803138018,
-0.09974778443574905,
0.02068343386054039,
0.036992646753787994,
0.06742747128009796,
0.000717418675776571,
-0.005685705225914717,
0.010919632390141487,
-0.037063077092170715,
0.2197755128145218,
-0.0656241923570633,
-0.09239254891872406,
-0.06753179430961609,
0.23204971849918365,
0.033168643712997437,
-0.0024129999801516533,
0.006973756942898035,
-0.08736778795719147,
-0.0013908683322370052,
0.19793660938739777,
0.16433179378509521,
-0.11321097612380981,
-0.016688499599695206,
-0.010998896323144436,
-0.02705851010978222,
-0.06523360311985016,
0.15385043621063232,
0.08979447931051254,
0.021714648231863976,
-0.1268528252840042,
-0.055480100214481354,
-0.06969206035137177,
-0.026057081297039986,
-0.05149729549884796,
0.02210139110684395,
0.06985922902822495,
0.0067195408046245575,
-0.057038526982069016,
0.07440855354070663,
-0.06528406590223312,
-0.12959152460098267,
0.09217079728841782,
-0.17180024087429047,
-0.16003258526325226,
-0.005107810255140066,
0.13685229420661926,
-0.014969736337661743,
0.07864218950271606,
-0.055147841572761536,
0.0020822961814701557,
0.0350295789539814,
-0.04324221611022949,
-0.10012427717447281,
-0.12564751505851746,
0.11123327165842056,
-0.12719842791557312,
0.19970452785491943,
-0.03664578124880791,
0.11292441934347153,
0.10818856954574585,
0.047670528292655945,
-0.06689070165157318,
0.06018618866801262,
0.034086115658283234,
-0.13643108308315277,
-0.013344066217541695,
0.11392532289028168,
-0.034242887049913406,
0.022440044209361076,
0.012739570811390877,
-0.12658178806304932,
0.037606626749038696,
-0.03797672316431999,
-0.04187591001391411,
-0.0537186861038208,
-0.08944118767976761,
-0.07715018093585968,
0.08797813951969147,
0.20356523990631104,
-0.00788735132664442,
0.045661602169275284,
-0.08690144121646881,
0.03324709087610245,
0.0873311311006546,
0.04610328748822212,
-0.08746055513620377,
-0.2449345886707306,
0.02052149921655655,
0.13203203678131104,
-0.07753560692071915,
-0.21851412951946259,
-0.08490725606679916,
0.026690108701586723,
-0.08245307207107544,
-0.07546021789312363,
0.0651690661907196,
0.10766445845365524,
0.06770363450050354,
-0.0422687791287899,
-0.14003241062164307,
-0.09490946680307388,
0.14470531046390533,
-0.14743778109550476,
-0.10100451856851578
] |
null | null | diffusers |
# LoRA DreamBooth - nick1221/outputs
These are LoRA adaption weights for stabilityai/stable-diffusion-2-1. The weights were trained on a photo of robot dog using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following.




LoRA for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "lora"], "base_model": "stabilityai/stable-diffusion-2-1", "instance_prompt": "a photo of robot dog", "inference": true} | text-to-image | nick1221/outputs | [
"diffusers",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"lora",
"base_model:stabilityai/stable-diffusion-2-1",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-06T04:35:35+00:00 | [] | [] | TAGS
#diffusers #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-stabilityai/stable-diffusion-2-1 #license-creativeml-openrail-m #region-us
|
# LoRA DreamBooth - nick1221/outputs
These are LoRA adaption weights for stabilityai/stable-diffusion-2-1. The weights were trained on a photo of robot dog using DreamBooth. You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
LoRA for the text encoder was enabled: False.
| [
"# LoRA DreamBooth - nick1221/outputs\n\nThese are LoRA adaption weights for stabilityai/stable-diffusion-2-1. The weights were trained on a photo of robot dog using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-stabilityai/stable-diffusion-2-1 #license-creativeml-openrail-m #region-us \n",
"# LoRA DreamBooth - nick1221/outputs\n\nThese are LoRA adaption weights for stabilityai/stable-diffusion-2-1. The weights were trained on a photo of robot dog using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
66,
95
] | [
"passage: TAGS\n#diffusers #stable-diffusion #stable-diffusion-diffusers #text-to-image #lora #base_model-stabilityai/stable-diffusion-2-1 #license-creativeml-openrail-m #region-us \n# LoRA DreamBooth - nick1221/outputs\n\nThese are LoRA adaption weights for stabilityai/stable-diffusion-2-1. The weights were trained on a photo of robot dog using DreamBooth. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3\n\n\nLoRA for the text encoder was enabled: False."
] | [
-0.011676653288304806,
0.006012909114360809,
-0.0020474332850426435,
0.02847246453166008,
0.13854143023490906,
-0.0036198622547090054,
0.22681288421154022,
0.08492279797792435,
0.045761313289403915,
0.07900574058294296,
0.0826135203242302,
0.06975211948156357,
-0.005930464714765549,
0.14287935197353363,
-0.004946372006088495,
-0.1460188925266266,
0.06048252806067467,
-0.08777972310781479,
-0.056101031601428986,
0.011286484077572823,
0.050094395875930786,
-0.02348863147199154,
0.0921449065208435,
-0.025634948164224625,
-0.14646904170513153,
0.036266278475522995,
-0.004645939450711012,
-0.042643506079912186,
0.05524274706840515,
0.06611350178718567,
0.01831171289086342,
0.056524284183979034,
0.05857698991894722,
-0.16211619973182678,
0.025475680828094482,
0.027221065014600754,
-0.03503834083676338,
0.04598848894238472,
-0.054308272898197174,
-0.038578346371650696,
0.12572255730628967,
-0.08467996120452881,
-0.021393386647105217,
0.015807252377271652,
-0.053287163376808167,
-0.021223871037364006,
-0.007526477333158255,
0.0036215640138834715,
0.040274329483509064,
0.10069236904382706,
-0.0208426583558321,
0.013071403838694096,
0.020364724099636078,
0.0696302205324173,
0.26351720094680786,
-0.22994261980056763,
-0.021011140197515488,
0.17708580195903778,
-0.04535144194960594,
0.11613206565380096,
-0.00684313103556633,
0.08907348662614822,
0.11468006670475006,
-0.049664463847875595,
-0.052337780594825745,
-0.026381397619843483,
0.01857505552470684,
-0.10773258656263351,
-0.08181434869766235,
0.00272190710529685,
0.17865248024463654,
-0.015865908935666084,
-0.09521862119436264,
-0.1357811689376831,
-0.022923963144421577,
0.11003577709197998,
-0.013996200636029243,
0.03825303167104721,
-0.018907466903328896,
0.022059256210923195,
-0.03442973271012306,
-0.06631210446357727,
-0.03322506323456764,
-0.10151402652263641,
-0.03832930326461792,
0.2825145423412323,
0.03266335278749466,
0.061832986772060394,
-0.013861238956451416,
0.15778188407421112,
-0.09019399434328079,
-0.11443639546632767,
0.03708569332957268,
-0.00542826484888792,
0.011950742453336716,
0.0918334349989891,
-0.058674298226833344,
-0.1064082607626915,
0.1051078587770462,
-0.015123462304472923,
0.04914277419447899,
-0.0065546599216759205,
-0.005150447599589825,
0.074192114174366,
-0.014865890145301819,
0.019808335229754448,
-0.09857050329446793,
-0.011764791794121265,
0.02181907370686531,
0.032259222120046616,
0.08381380885839462,
-0.057372499257326126,
-0.13272833824157715,
0.007094253320246935,
-0.04948872700333595,
0.03781973943114281,
0.001094129285775125,
0.001806468004360795,
-0.07818447053432465,
0.009868742898106575,
0.005233233328908682,
-0.024655954912304878,
-0.025077778846025467,
-0.05733119323849678,
-0.041503239423036575,
0.0634949803352356,
0.11681912839412689,
0.04107257351279259,
0.0006062744068913162,
0.032148998230695724,
-0.03611922264099121,
0.032434288412332535,
-0.024873681366443634,
-0.11177108436822891,
-0.030310021713376045,
-0.08896451443433762,
0.028140058740973473,
-0.15906774997711182,
-0.09007349610328674,
0.021461008116602898,
-0.0019751591607928276,
-0.07639370113611221,
0.052887529134750366,
-0.08426635712385178,
-0.12158353626728058,
-0.001248635700903833,
0.04608866572380066,
-0.017462244257330894,
-0.03673672303557396,
0.06952572613954544,
0.04609818756580353,
0.14167729020118713,
-0.1593647301197052,
-0.030299479141831398,
-0.08705614507198334,
0.006854345090687275,
-0.05528628081083298,
0.13765327632427216,
-0.015469810925424099,
0.03514067456126213,
-0.06439682096242905,
0.0010783018078655005,
-0.01922229304909706,
0.03678669035434723,
0.06470558792352676,
0.11198557168245316,
-0.20500598847866058,
-0.04550372436642647,
0.1194082573056221,
-0.1763269305229187,
-0.0874277725815773,
0.0851345881819725,
0.011961463838815689,
0.1421789526939392,
0.08733077347278595,
0.06694456934928894,
0.019853731617331505,
-0.2551993131637573,
-0.04373886063694954,
-0.10106795281171799,
-0.048199206590652466,
0.026972057297825813,
0.00999737624078989,
0.08142144978046417,
-0.08063264936208725,
0.0510840080678463,
-0.11595971882343292,
0.11071451008319855,
-0.05767287686467171,
-0.01875344105064869,
-0.03746822848916054,
-0.07001993060112,
0.01487185712903738,
-0.00043462993926368654,
0.07942695170640945,
-0.011508272960782051,
-0.04461773484945297,
0.010191122069954872,
0.058274559676647186,
-0.015156475827097893,
0.0042624399065971375,
-0.03654301539063454,
0.11923305690288544,
-0.09033119678497314,
0.02307468093931675,
-0.11060400307178497,
-0.00917879305779934,
0.026814142242074013,
0.19072438776493073,
0.13160113990306854,
0.05051349475979805,
0.09027320146560669,
0.06907682865858078,
-0.0016890146071091294,
-0.03444882854819298,
0.04797802120447159,
0.004302643705159426,
-0.051186371594667435,
-0.14482657611370087,
0.09205939620733261,
-0.09501803666353226,
0.08187990635633469,
-0.11202023923397064,
0.03469770401716232,
-0.02501971833407879,
0.17211535573005676,
0.07600358128547668,
-0.016379013657569885,
0.0334969162940979,
0.05069991573691368,
-0.08460938930511475,
-0.04312046617269516,
0.008143371902406216,
-0.013605413027107716,
-0.16590793430805206,
0.1697826385498047,
-0.15405458211898804,
0.012434735894203186,
0.11634813994169235,
0.09891583025455475,
-0.043320201337337494,
-0.01011803187429905,
0.023336708545684814,
0.03752376139163971,
-0.05055709555745125,
-0.056561291217803955,
0.1458178013563156,
-0.015385990031063557,
0.13719190657138824,
-0.022267818450927734,
0.09850627928972244,
0.041689854115247726,
-0.10413670539855957,
-0.08756818622350693,
0.06405187398195267,
-0.05304389074444771,
0.05489198490977287,
0.004364029038697481,
0.050056342035532,
-0.02904283069074154,
0.1795942783355713,
0.01771496795117855,
-0.054786503314971924,
-0.058059338480234146,
0.061375442892313004,
0.06238449737429619,
0.07685752958059311,
0.07431969791650772,
-0.005667429883033037,
-0.00590005237609148,
-0.03281088173389435,
0.010480673052370548,
-0.16510500013828278,
-0.022168375551700592,
0.004879968706518412,
-0.07571558654308319,
0.1615661233663559,
0.06883100420236588,
-0.07573951035737991,
0.08838073164224625,
-0.07019739598035812,
-0.11258729547262192,
0.01019972562789917,
-0.024538470432162285,
-0.03143394738435745,
0.07691542059183121,
-0.06887198239564896,
-0.16837751865386963,
-0.18726272881031036,
-0.006261173170059919,
-0.04173015058040619,
0.01209690235555172,
0.0656493529677391,
-0.11308588087558746,
-0.0242390725761652,
-0.09178086370229721,
0.08302127569913864,
0.033484023064374924,
0.04502033814787865,
0.10195645689964294,
0.0027631332632154226,
0.02687872014939785,
-0.02186974138021469,
-0.03539779409766197,
-0.057602282613515854,
0.041183192282915115,
0.05484450235962868,
0.0045309122651815414,
0.11681906878948212,
0.0694626197218895,
0.02162833884358406,
0.01451404020190239,
0.01872984878718853,
0.20824579894542694,
-0.017947092652320862,
0.004942829720675945,
0.17159639298915863,
0.06310028582811356,
0.04573884233832359,
0.1416800320148468,
0.014090711250901222,
-0.1049753949046135,
0.10717226564884186,
-0.007133753504604101,
-0.14505137503147125,
-0.035434216260910034,
-0.0983230471611023,
-0.020256806164979935,
-0.08563357591629028,
0.02288808673620224,
0.04178376495838165,
0.09728863835334778,
0.07926411926746368,
0.053043924272060394,
0.021161211654543877,
0.0477512963116169,
0.07235304266214371,
0.015017923898994923,
-0.056953899562358856,
0.06332813948392868,
-0.06681250780820847,
-0.12130957096815109,
0.09629011154174805,
-0.06575050950050354,
0.18033155798912048,
-0.0779665932059288,
-0.0087777990847826,
0.03325730934739113,
-0.025536173954606056,
0.06837107986211777,
0.027372559532523155,
-0.006723945029079914,
-0.0373404361307621,
-0.03042260929942131,
-0.1147598996758461,
0.05291488394141197,
0.1272348314523697,
0.03206713870167732,
0.012223350815474987,
-0.030507156625390053,
0.1042146161198616,
0.01834738999605179,
0.07526528835296631,
0.14641356468200684,
-0.2629661560058594,
0.022152140736579895,
0.0019399382872506976,
0.0318424329161644,
0.013023564592003822,
-0.014746333472430706,
0.23079879581928253,
-0.023134667426347733,
0.012471770867705345,
-0.035622142255306244,
0.031306762248277664,
-0.013804427348077297,
-0.03737487643957138,
-0.09847963601350784,
0.1506483554840088,
-0.026181520894169807,
-0.031155942007899284,
-0.2226674109697342,
0.056226614862680435,
-0.01058986596763134,
0.03543202579021454,
-0.03057287074625492,
-0.00016637498629279435,
0.03851735591888428,
-0.018216049298644066,
0.06714294850826263,
-0.006700744852423668,
-0.03210629150271416,
-0.1659415066242218,
-0.17076224088668823,
-0.0257611945271492,
0.07253918796777725,
-0.03845297917723656,
0.08343641459941864,
0.031116612255573273,
0.015019717626273632,
-0.0021596767473965883,
0.034603968262672424,
-0.1472705900669098,
-0.09569238871335983,
-0.015140075236558914,
0.22122225165367126,
0.07210521399974823,
-0.07092747837305069,
-0.08702176064252853,
0.044544484466314316,
0.0816856324672699,
-0.00027893114020116627,
-0.0836513489484787,
-0.05391190946102142,
0.012300430797040462,
0.054423026740550995,
-0.034766215831041336,
-0.03059522435069084,
-0.04931925982236862,
0.025905728340148926,
-0.06317928433418274,
-0.14253032207489014,
0.014773698523640633,
-0.04834126681089401,
-0.139130637049675,
-0.07211072742938995,
0.03848591074347496,
0.01817358285188675,
-0.0013770643854513764,
-0.0030409761238843203,
0.020114365965127945,
0.037835948169231415,
-0.058587126433849335,
0.010859607718884945,
0.2125176638364792,
-0.15869073569774628,
0.17059801518917084,
-0.011202823370695114,
-0.011910845525562763,
-0.10844641923904419,
-0.01351979747414589,
0.1130821704864502,
0.2628784477710724,
-0.07038748264312744,
0.06536073982715607,
0.07812145352363586,
-0.06951749324798584,
-0.19047392904758453,
-0.11416029185056686,
0.012570417486131191,
0.041750919073820114,
-0.017894214019179344,
-0.09622295200824738,
0.09497074037790298,
0.03925476223230362,
-0.00010226306039839983,
0.16109196841716766,
-0.37921589612960815,
-0.1297871172428131,
-0.0029549847822636366,
0.1617010086774826,
0.3175964057445526,
-0.14384658634662628,
-0.053403791040182114,
-0.018133506178855896,
-0.03653629124164581,
0.08411423861980438,
0.017858989536762238,
0.13634391129016876,
-0.028045469895005226,
-0.0031166065018624067,
0.04123726487159729,
-0.008927840739488602,
0.15672533214092255,
-0.05461661145091057,
0.016009513288736343,
-0.08465912193059921,
-0.007066451478749514,
0.06609058380126953,
-0.05378445237874985,
0.056754447519779205,
-0.20398737490177155,
0.02308882586658001,
-0.04224129021167755,
-0.008413758128881454,
0.022423122078180313,
-0.03533649072051048,
0.02218104898929596,
-0.0755867287516594,
-0.09657033532857895,
0.037150293588638306,
0.012125087901949883,
-0.011973969638347626,
0.03873773291707039,
-0.01997818425297737,
-0.01932232268154621,
0.1791345477104187,
-0.10191982984542847,
0.07030462473630905,
0.023158658295869827,
-0.00690061878412962,
-0.041471973061561584,
0.11785146594047546,
-0.09296199679374695,
-0.015134166926145554,
0.14013421535491943,
0.059278205037117004,
0.1222592294216156,
0.004469454754143953,
-0.11135589331388474,
0.10947614908218384,
0.12462295591831207,
-0.07375124096870422,
-0.08687618374824524,
-0.017568299546837807,
0.029314963147044182,
0.11647491157054901,
-0.03447107598185539,
0.21122193336486816,
-0.10080781579017639,
0.0505848154425621,
-0.03087017685174942,
0.02422158420085907,
-0.04637204855680466,
0.12096612900495529,
-0.05069602653384209,
0.008606821298599243,
-0.06342050433158875,
0.08319738507270813,
-0.02235689014196396,
0.06009584665298462,
0.06703609973192215,
-0.004516317509114742,
-0.05128618702292442,
0.026872806251049042,
0.015975236892700195,
0.14866842329502106,
-0.12127729505300522,
-0.047217629849910736,
-0.16320998966693878,
-0.10233759880065918,
-0.0067615206353366375,
0.03944955766201019,
0.04714512079954147,
0.01741105318069458,
-0.03768639266490936,
-0.04104430228471756,
-0.052349384874105453,
0.02902177721261978,
0.06821791082620621,
0.059970613569021225,
-0.24428491294384003,
-0.03936731442809105,
0.014192759059369564,
-0.01895584724843502,
-0.08301818370819092,
-0.08354833722114563,
-0.11745733767747879,
0.002019972074776888,
-0.12857185304164886,
0.09271793812513351,
-0.05951184406876564,
-0.05754411965608597,
0.009846705943346024,
-0.03518904000520706,
0.013033885508775711,
0.03119347058236599,
-0.0020794214215129614,
-0.024215560406446457,
-0.014141635969281197,
-0.0102777611464262,
-0.03899579495191574,
-0.093644879758358,
-0.03781772777438164,
-0.06973141431808472,
0.01597319170832634,
0.04308704659342766,
-0.06916506588459015,
0.012943267822265625,
-0.2359122484922409,
0.05279616266489029,
0.20527960360050201,
-0.03165151923894882,
-0.00015732452447991818,
-0.06775474548339844,
0.023161105811595917,
-0.009617610834538937,
0.05091432109475136,
0.011987065896391869,
0.0676177442073822,
-0.06249356269836426,
-0.03406140208244324,
-0.06071053072810173,
0.02948548085987568,
-0.044844187796115875,
0.04679473862051964,
0.17963214218616486,
0.1488623321056366,
0.15992332994937897,
-0.1658790409564972,
0.03847486525774002,
-0.025129584595561028,
-0.009825136512517929,
-0.016188232228159904,
-0.021325793117284775,
0.04219816252589226,
-0.06883878260850906,
-0.02490762434899807,
0.00835413672029972,
0.09848900884389877,
0.013478117063641548,
-0.18584609031677246,
-0.014543125405907631,
0.01451915130019188,
0.007560160476714373,
-0.01357716042548418,
0.1704014390707016,
0.03949255123734474,
0.0522591732442379,
-0.1279767006635666,
0.048460569232702255,
0.17365317046642303,
0.1066819429397583,
0.060295458883047104,
0.07740763574838638,
0.06605257838964462,
0.10613496601581573,
0.07276079803705215,
0.050676919519901276,
0.0011516428785398602,
0.08519910275936127,
-0.11485487222671509,
0.07098987698554993,
-0.04904291033744812,
-0.009650930762290955,
0.11437734216451645,
-0.03526164963841438,
-0.041702598333358765,
0.08796532452106476,
-0.0028013354167342186,
-0.04272100329399109,
-0.06123891472816467,
-0.05515005439519882,
-0.1931755393743515,
0.01692485436797142,
-0.03745332732796669,
0.017998099327087402,
0.05781655013561249,
0.03771117702126503,
0.03501157835125923,
0.06557292491197586,
-0.00418962724506855,
-0.05503072962164879,
0.08253085613250732,
-0.05383099243044853,
-0.07772582024335861,
0.06661383807659149,
-0.034735776484012604,
0.08399248868227005,
0.03348363935947418,
-0.016817299649119377,
0.11438742280006409,
0.016411039978265762,
0.03396404907107353,
-0.03568677604198456,
-0.07073526084423065,
-0.002068187342956662,
0.01942046545445919,
0.032965585589408875,
0.1803436428308487,
0.1389022171497345,
-0.07723056524991989,
-0.046238481998443604,
0.20061077177524567,
-0.060491498559713364,
-0.01847691833972931,
-0.16271957755088806,
0.16722941398620605,
-0.1378687471151352,
0.06157577410340309,
-0.06005355343222618,
-0.11308230459690094,
-0.11458934098482132,
0.10227760672569275,
0.10460751503705978,
-0.09270686656236649,
0.008660721592605114,
-0.11196740716695786,
-0.003213996533304453,
-0.049098819494247437,
0.07695260643959045,
0.05143670737743378,
0.23518043756484985,
-0.002241021255031228,
0.03221474587917328,
-0.11572054028511047,
-0.08472517132759094,
-0.07696788012981415,
-0.12636414170265198,
-0.038566939532756805,
-0.02603863552212715,
-0.09303196519613266,
0.047203898429870605,
-0.14461112022399902,
-0.1487552970647812,
0.19192028045654297,
-0.12965823709964752,
-0.020810434594750404,
-0.04595347121357918,
0.02403257042169571,
0.05586421117186546,
0.03565829619765282,
-0.08163128793239594,
0.027558231726288795,
-0.030525123700499535,
-0.03577347844839096,
-0.06990934908390045,
-0.01699887588620186,
-0.08426602929830551,
-0.15350256860256195,
0.08303628116846085,
-0.023500440642237663,
0.02482873760163784,
0.048125676810741425,
0.012027431279420853,
-0.11363935470581055,
0.1265796422958374,
-0.08701340854167938,
-0.07887069880962372,
-0.07158258557319641,
0.155805766582489,
-0.02993597462773323,
0.10858611017465591,
0.06821929663419724,
-0.06039667874574661,
0.01188435684889555,
0.0695423036813736,
-0.03927082568407059,
-0.1280103474855423,
-0.01723094843327999,
-0.07516665756702423,
0.08267499506473541,
0.11218932271003723,
-0.035789232701063156,
0.05841498076915741,
0.03157120943069458,
0.030842043459415436,
0.01704878732562065,
-0.05041337013244629,
0.023399269208312035,
-0.12361639738082886,
-0.0203224029392004,
0.030660582706332207,
0.018037481233477592,
-0.32243645191192627,
-0.08854703605175018,
-0.16974565386772156,
-0.013632378540933132,
0.021596817299723625,
0.09850535541772842,
0.2777039408683777,
0.007117736618965864,
0.009741304442286491,
-0.2615061402320862,
0.009785203263163567,
0.10829973220825195,
-0.045041345059871674,
-0.08665109425783157
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-large-bn-adapter-3.17M-snli-model1
This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6213
- Accuracy: 0.804
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 15
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.3105 | 1.0 | 8584 | 0.2418 | 0.9176 |
| 0.2877 | 2.0 | 17168 | 0.2294 | 0.9223 |
| 0.2768 | 3.0 | 25752 | 0.2243 | 0.9240 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "roberta-large", "model-index": [{"name": "roberta-large-bn-adapter-3.17M-snli-model1", "results": []}]} | null | varun-v-rao/roberta-large-bn-adapter-3.17M-snli-model1 | [
"tensorboard",
"generated_from_trainer",
"base_model:roberta-large",
"license:mit",
"region:us"
] | 2024-02-06T04:35:56+00:00 | [] | [] | TAGS
#tensorboard #generated_from_trainer #base_model-roberta-large #license-mit #region-us
| roberta-large-bn-adapter-3.17M-snli-model1
==========================================
This model is a fine-tuned version of roberta-large on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6213
* Accuracy: 0.804
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 15
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 15\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-large #license-mit #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 15\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
32,
98,
4,
33
] | [
"passage: TAGS\n#tensorboard #generated_from_trainer #base_model-roberta-large #license-mit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 15\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.0771183967590332,
-0.02338598109781742,
-0.0009387187310494483,
0.11565724015235901,
0.19629453122615814,
0.03765023872256279,
0.15827122330665588,
0.0669541135430336,
-0.1202378049492836,
0.016902709379792213,
0.1015181913971901,
0.12441790103912354,
-0.006471141707152128,
0.05422975867986679,
-0.02366012893617153,
-0.2654603123664856,
-0.03600674867630005,
0.042314834892749786,
-0.13900874555110931,
0.12058179825544357,
0.0820549950003624,
-0.14838092029094696,
0.06784864515066147,
0.007021482102572918,
-0.3299749195575714,
0.05311701446771622,
0.040150344371795654,
-0.04650535061955452,
0.13689495623111725,
-0.005959800910204649,
0.1912897527217865,
-0.014357205480337143,
0.13535788655281067,
-0.12236494570970535,
0.012554151937365532,
0.0692172572016716,
0.0033135872799903154,
0.0880010724067688,
0.0692153051495552,
0.03138509765267372,
0.1041935458779335,
-0.08331011235713959,
0.062019456177949905,
0.018592795357108116,
-0.13079258799552917,
-0.25760480761528015,
-0.06458064913749695,
-0.012118471786379814,
0.04803662747144699,
0.08139725774526596,
-0.008210346102714539,
0.1446983963251114,
-0.0920000672340393,
0.08780413866043091,
0.26491448283195496,
-0.25538280606269836,
-0.08071357756853104,
0.08349694311618805,
0.002864832291379571,
0.09621281921863556,
-0.11272162944078445,
-0.025132393464446068,
0.07316576689481735,
0.044374264776706696,
0.1275271773338318,
-0.027044352144002914,
-0.12679032981395721,
0.04078172892332077,
-0.14756456017494202,
0.01662489026784897,
0.06265874207019806,
0.017955690622329712,
-0.01668640226125717,
0.03745260462164879,
-0.08426281064748764,
-0.1387530267238617,
-0.05932510644197464,
-0.022136179730296135,
0.08716170489788055,
-0.05317554250359535,
-0.13433724641799927,
-0.03702313080430031,
-0.10672880709171295,
-0.10602474957704544,
-0.06517684459686279,
0.21792365610599518,
0.052122946828603745,
0.04522370547056198,
-0.07559802383184433,
0.10595238208770752,
-0.05941036716103554,
-0.13071104884147644,
0.06065317243337631,
0.04801686480641365,
-0.01854735054075718,
-0.05604678392410278,
-0.08025732636451721,
-0.10938850045204163,
0.010126997716724873,
0.10087014734745026,
-0.09256834536790848,
0.04306581988930702,
0.05365421250462532,
0.06441201269626617,
-0.10520296543836594,
0.18004430830478668,
-0.05957877263426781,
0.03857171908020973,
0.00844933744519949,
0.029741238802671432,
0.0030724024400115013,
0.0017921003745868802,
-0.12343105673789978,
0.01895955391228199,
0.07241465151309967,
-0.008542525582015514,
-0.08826280385255814,
0.035553738474845886,
-0.02374984510242939,
-0.0019674485083669424,
-0.02712722308933735,
-0.08307935297489166,
0.052716758102178574,
-0.0037567485123872757,
-0.09891849756240845,
-0.0084174619987607,
0.007129758130759001,
0.008957339450716972,
0.027052106335759163,
0.15098656713962555,
-0.11759382486343384,
0.08525898307561874,
-0.14702323079109192,
-0.12383737415075302,
0.0025385068729519844,
-0.036478884518146515,
0.026542475447058678,
-0.07089141756296158,
-0.13911156356334686,
-0.004245743155479431,
0.06496331840753555,
-0.04460037499666214,
-0.005914242938160896,
-0.025369374081492424,
-0.0991542637348175,
-0.009822631254792213,
-0.01434100978076458,
0.18902897834777832,
-0.04149995744228363,
0.10451147705316544,
0.05649450421333313,
0.06608366966247559,
-0.11581817269325256,
0.046223487704992294,
-0.09225020557641983,
0.0035877241753041744,
-0.27090752124786377,
0.007403554394841194,
-0.05449469015002251,
0.056463830173015594,
-0.05611887574195862,
-0.10521257668733597,
0.011278300546109676,
0.004109518602490425,
0.109855517745018,
0.06188524141907692,
-0.19664910435676575,
-0.08374259620904922,
0.1265738159418106,
-0.07208211719989777,
-0.06966347992420197,
0.09658443182706833,
-0.0675402283668518,
0.0616341158747673,
0.06034306064248085,
0.19330619275569916,
-0.04018179327249527,
-0.1117883026599884,
0.031121429055929184,
-0.02538968063890934,
0.03616293519735336,
-0.10308338701725006,
0.040563926100730896,
-0.01234069000929594,
0.02253156527876854,
0.028139596804976463,
-0.04745751619338989,
0.06662216782569885,
-0.12554529309272766,
-0.07301375269889832,
-0.0532236211001873,
-0.09742984175682068,
0.01884237863123417,
0.08304493129253387,
0.10052715986967087,
-0.1001797541975975,
-0.05883792042732239,
0.20200799405574799,
0.05182719975709915,
-0.055199600756168365,
0.027395352721214294,
-0.0374019593000412,
0.05941452458500862,
-0.08535128086805344,
-0.0429525189101696,
-0.19756273925304413,
-0.06119018793106079,
0.00355530739761889,
0.0018121612956747413,
0.07383105158805847,
0.03060845658183098,
0.09175087511539459,
0.07822085916996002,
-0.07687509059906006,
0.01908501796424389,
-0.08148499578237534,
-0.009355555288493633,
-0.14051076769828796,
-0.21513210237026215,
-0.04316600412130356,
-0.007547440472990274,
0.05666517838835716,
-0.19102701544761658,
0.02760765515267849,
-0.040894560515880585,
0.07800427079200745,
0.009341584518551826,
-0.020608922466635704,
-0.07544775307178497,
0.10123533755540848,
0.0028144458774477243,
-0.06340022385120392,
0.04522160068154335,
-0.01900072768330574,
-0.060669369995594025,
-0.08676217496395111,
-0.08663681894540787,
0.18203702569007874,
0.14117646217346191,
-0.1575946807861328,
-0.09752177447080612,
0.029167119413614273,
-0.06841655820608139,
-0.02319113165140152,
-0.06353634595870972,
0.05798840522766113,
0.17781010270118713,
-0.01677885092794895,
0.1393747478723526,
-0.08518370985984802,
-0.04166678711771965,
0.003874078392982483,
-0.04113404080271721,
0.06541948765516281,
0.08322487026453018,
0.19088082015514374,
-0.05349712073802948,
0.11812462657690048,
0.14811624586582184,
-0.16714049875736237,
0.13760140538215637,
-0.05634993314743042,
-0.08346540480852127,
-0.037074003368616104,
-0.03322187811136246,
-0.010251283645629883,
0.1673094630241394,
-0.06681815534830093,
0.008840254507958889,
-0.006052156910300255,
0.00826533604413271,
0.04873286932706833,
-0.25399911403656006,
-0.05493907630443573,
0.003615438938140869,
-0.031109502539038658,
-0.01897435635328293,
-0.034361831843853,
0.01771344244480133,
0.10878857225179672,
-0.03234705701470375,
-0.07580333203077316,
0.019361982122063637,
0.01726396568119526,
-0.05085916444659233,
0.2163008451461792,
-0.06027424708008766,
-0.008800537325441837,
-0.110052190721035,
-0.03382859379053116,
-0.055247966200113297,
-0.011952687986195087,
0.030048366636037827,
-0.13012148439884186,
-0.019391417503356934,
-0.03154926374554634,
0.05755433067679405,
0.00856607872992754,
0.04654748737812042,
-0.016443248838186264,
0.007641960866749287,
0.08074817061424255,
-0.13844440877437592,
0.019007697701454163,
-0.09597563743591309,
-0.09688852727413177,
0.02764432318508625,
0.08910829573869705,
0.12559741735458374,
0.17223350703716278,
-0.04191087558865547,
-0.00016152928583323956,
-0.010333864949643612,
0.26281681656837463,
-0.06582499295473099,
-0.0409972220659256,
0.09836961328983307,
0.0040198699571192265,
0.05359918251633644,
0.09495239704847336,
0.12156028300523758,
-0.10916731506586075,
-0.011763069778680801,
0.05471184849739075,
-0.05523824319243431,
-0.22525060176849365,
-0.015372201800346375,
-0.026568099856376648,
-0.09303747862577438,
0.061492808163166046,
0.03546183928847313,
0.011672494933009148,
0.0626130998134613,
0.07001728564500809,
0.11822250485420227,
-0.07602109760046005,
0.044161152094602585,
0.05650758370757103,
0.06062876060605049,
0.11701733618974686,
-0.049334123730659485,
-0.08709938824176788,
0.017035862430930138,
-0.02792544662952423,
0.2608940899372101,
-0.003234792035073042,
0.11262992769479752,
0.0704139694571495,
0.17441707849502563,
0.0013825366040691733,
0.09297072887420654,
0.002681716112419963,
-0.06783977150917053,
-0.0010283893207088113,
-0.0514882393181324,
0.03036908060312271,
0.008914492093026638,
-0.06930924952030182,
0.05557221546769142,
-0.07630528509616852,
-0.014072875492274761,
0.054463934153318405,
0.22022049129009247,
-0.006893052253872156,
-0.30944252014160156,
-0.04415511339902878,
-0.015390132553875446,
0.0022743132431060076,
-0.004439713899046183,
0.0007083883392624557,
0.15097948908805847,
-0.035689547657966614,
-0.001854890724644065,
-0.06485070288181305,
0.08073244243860245,
-0.0021306131966412067,
0.030154330655932426,
0.08110545575618744,
0.18307706713676453,
-0.01473045814782381,
0.04507400467991829,
-0.28625962138175964,
0.2966838479042053,
0.031610965728759766,
0.10888472199440002,
-0.04415631666779518,
-0.015808187425136566,
0.016955560073256493,
0.024775445461273193,
0.00192355050239712,
-0.02387166954576969,
-0.03413400053977966,
-0.20917373895645142,
-0.031920287758111954,
0.05079735442996025,
0.13690678775310516,
0.013592569157481194,
0.08300565928220749,
0.003621449461206794,
0.011285382322967052,
0.10256077349185944,
-0.061193015426397324,
-0.10951033979654312,
-0.03771475329995155,
-0.06412900239229202,
0.00858460646122694,
-0.11919497698545456,
-0.05595990642905235,
-0.11280938982963562,
-0.09785408526659012,
0.10614145547151566,
0.03506643697619438,
-0.029993409290909767,
-0.12049738317728043,
0.13961537182331085,
0.10508313030004501,
-0.06063486263155937,
0.04410893842577934,
0.02173878252506256,
0.04964642971754074,
0.03310230001807213,
-0.06378598511219025,
0.1224171370267868,
-0.042965106666088104,
-0.14141085743904114,
-0.049504365772008896,
0.04634576290845871,
0.04799007251858711,
0.06201613321900368,
-0.03265991061925888,
0.016433600336313248,
-0.009306224063038826,
-0.10899034887552261,
0.025896292179822922,
-0.06568954885005951,
0.027306046336889267,
0.006993108429014683,
-0.03399878367781639,
0.04454728588461876,
-0.032386358827352524,
0.01154983602464199,
0.12019670754671097,
0.27254003286361694,
-0.08134222030639648,
-0.03322385624051094,
0.04883117228746414,
-0.0625927671790123,
-0.16976670920848846,
0.1174684539437294,
0.07636942714452744,
-0.011463976465165615,
0.0691247433423996,
-0.16874659061431885,
0.16138216853141785,
0.1374076008796692,
-0.014330890960991383,
0.12130691856145859,
-0.3171677887439728,
-0.15192899107933044,
0.07728830724954605,
0.2083645462989807,
0.1414853036403656,
-0.15298859775066376,
-0.018538478761911392,
-0.024304265156388283,
-0.08696804940700531,
0.11098819971084595,
-0.149994894862175,
0.10798642784357071,
-0.014368582516908646,
0.10877203196287155,
-0.0075237383134663105,
-0.07077398896217346,
0.11627230793237686,
0.020119724795222282,
0.14833849668502808,
-0.036858201026916504,
-0.030050218105316162,
0.11731382459402084,
-0.00822464656084776,
-0.002923825057223439,
-0.021162990480661392,
0.0065061431378126144,
-0.007424949202686548,
-0.005688135512173176,
-0.08801580965518951,
0.04691499471664429,
-0.03632216900587082,
-0.07162144780158997,
-0.05233423411846161,
0.0005981061840429902,
-0.0247624721378088,
-0.032209575176239014,
0.13083720207214355,
0.058521151542663574,
0.14990085363388062,
0.09069039672613144,
0.012849832884967327,
-0.06014116853475571,
-0.07597269862890244,
0.022783732041716576,
-0.008354360237717628,
0.08113815635442734,
-0.17767250537872314,
-0.012214570306241512,
0.12224103510379791,
0.04019548371434212,
0.08837991952896118,
0.10206737369298935,
-0.05905798077583313,
0.03693832829594612,
0.07295778393745422,
-0.1445906162261963,
-0.10667029023170471,
0.020636167377233505,
-0.10017073154449463,
-0.07669620215892792,
0.06623666733503342,
0.07189833372831345,
-0.08332093805074692,
-0.001494182855822146,
-0.006749847438186407,
-0.03506280481815338,
-0.08269495517015457,
0.21886414289474487,
0.11235162615776062,
0.033901672810316086,
-0.1142057254910469,
0.07815881073474884,
0.0524623729288578,
-0.026110315695405006,
-0.006373128853738308,
0.06996048241853714,
-0.03993303328752518,
0.015185813419520855,
0.1417478621006012,
0.23321528732776642,
-0.1012658104300499,
-0.03262137621641159,
-0.1823062300682068,
-0.08904610574245453,
0.03994755074381828,
0.17964445054531097,
0.10788845270872116,
-0.025356709957122803,
-0.03256720304489136,
0.04756008833646774,
-0.15640318393707275,
0.05812344327569008,
0.0010129042202606797,
0.09082172811031342,
-0.12202759832143784,
0.1842188835144043,
0.027142224833369255,
0.02394787408411503,
-0.03145278990268707,
0.05900014936923981,
-0.14246414601802826,
0.025762224569916725,
-0.11849690973758698,
-0.06959079205989838,
0.024210864678025246,
-0.016295837238430977,
-0.0015753447078168392,
-0.07224774360656738,
-0.08365067839622498,
0.02747107669711113,
-0.14645402133464813,
-0.016882790252566338,
0.06349019706249237,
0.012220742180943489,
-0.14983715116977692,
-0.011154254898428917,
0.018556220456957817,
-0.041607361286878586,
0.02616596221923828,
0.05130140483379364,
0.03700418770313263,
0.09631823748350143,
-0.1948181539773941,
-0.01622881181538105,
0.06958498805761337,
-0.010168914683163166,
0.10178624838590622,
0.00007164372800616547,
-0.02618023380637169,
-0.027320828288793564,
0.14965637028217316,
0.006315896287560463,
0.0532679408788681,
-0.13455252349376678,
0.009545865468680859,
-0.05944815278053284,
-0.08920828998088837,
-0.049640290439128876,
0.005600042175501585,
0.07962608337402344,
0.030486276373267174,
0.15962567925453186,
-0.06881576031446457,
0.023795291781425476,
-0.22912809252738953,
-0.011988642625510693,
-0.005973231978714466,
-0.08066005259752274,
-0.09607843309640884,
-0.04385130852460861,
0.06769213825464249,
-0.05801203101873398,
0.14556249976158142,
0.0615113228559494,
0.04852116107940674,
0.03681236878037453,
0.025789150968194008,
0.005756925325840712,
0.029473168775439262,
0.21982179582118988,
0.032404184341430664,
-0.01600004732608795,
0.06612130254507065,
0.09754976630210876,
0.12349262088537216,
0.09205649048089981,
0.22340475022792816,
0.14668364822864532,
-0.06310050189495087,
0.09670358896255493,
0.07616374641656876,
-0.05525147169828415,
-0.12280090898275375,
0.08258674293756485,
-0.027110395953059196,
0.06667367368936539,
-0.0365634448826313,
0.20259621739387512,
0.07934638112783432,
-0.15638786554336548,
0.03500370681285858,
-0.034882865846157074,
-0.09940323233604431,
-0.10416051000356674,
0.03289220482110977,
-0.07379807531833649,
-0.17581431567668915,
0.03768274560570717,
-0.1002679392695427,
0.013114022091031075,
0.15000994503498077,
-0.005499658640474081,
-0.0005718793836422265,
0.1993098258972168,
0.07389441877603531,
0.05650206655263901,
0.017140554264187813,
-0.0008865993586368859,
-0.039375465363264084,
-0.09882354736328125,
-0.07234601676464081,
-0.02332313172519207,
-0.021074621006846428,
0.020581016317009926,
-0.056139301508665085,
-0.09612157940864563,
0.036306481808423996,
-0.0023770604748278856,
-0.09524495154619217,
0.01997782289981842,
0.03584783896803856,
0.07436777651309967,
-0.006363511085510254,
-0.009525413624942303,
0.011553548276424408,
-0.04284795746207237,
0.2112344205379486,
-0.0698825940489769,
-0.08222009986639023,
-0.06743564456701279,
0.22572340071201324,
0.019599953666329384,
-0.0018435685196891427,
0.011246087029576302,
-0.09043927490711212,
-0.010741928592324257,
0.19313791394233704,
0.1747453510761261,
-0.11656581610441208,
-0.015853330492973328,
-0.009869590401649475,
-0.028865229338407516,
-0.08068566024303436,
0.15508081018924713,
0.09705958515405655,
0.03811904415488243,
-0.12231825292110443,
-0.05177832022309303,
-0.06341677904129028,
-0.018754860386252403,
-0.04701407998800278,
0.03333538770675659,
0.0642240047454834,
0.010138324461877346,
-0.05823243036866188,
0.06959935277700424,
-0.06751707941293716,
-0.12411986291408539,
0.09368286281824112,
-0.16625095903873444,
-0.16437239944934845,
-0.0010534949833527207,
0.13753418624401093,
-0.0250567514449358,
0.08140779286623001,
-0.05507710203528404,
0.002505070995539427,
0.02866566739976406,
-0.04658779874444008,
-0.10609740763902664,
-0.1172671765089035,
0.10856945812702179,
-0.12338132411241531,
0.21073636412620544,
-0.036796264350414276,
0.11158949881792068,
0.10822150856256485,
0.05233068764209747,
-0.07007322460412979,
0.05466989800333977,
0.032422542572021484,
-0.13995181024074554,
-0.012045517563819885,
0.121559277176857,
-0.03499062359333038,
0.019619295373558998,
0.012560093775391579,
-0.12834806740283966,
0.03404403105378151,
-0.053176648914813995,
-0.04231993108987808,
-0.05712468922138214,
-0.09062554687261581,
-0.07597530633211136,
0.08623010665178299,
0.20477764308452606,
-0.0021313270553946495,
0.03720341995358467,
-0.0800006166100502,
0.03308482468128204,
0.08793040364980698,
0.048911623656749725,
-0.07967725396156311,
-0.24482928216457367,
0.021902302280068398,
0.1300877332687378,
-0.07942686975002289,
-0.22384104132652283,
-0.07741563767194748,
0.0341048426926136,
-0.07970941066741943,
-0.07915681600570679,
0.05733640491962433,
0.1190863698720932,
0.06894492357969284,
-0.04461020976305008,
-0.13503792881965637,
-0.09932435303926468,
0.1374901831150055,
-0.1596968173980713,
-0.09592102468013763
] |
null | null | diffusers |
# ae-real-nigi3d API Inference

## Get API Key
Get API key from [ModelsLab API](http://modelslab.com), No Payment needed.
Replace Key in below code, change **model_id** to "ae-real-nigi3d"
Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://modelslab.com/docs)
Try model for free: [Generate Images](https://modelslab.com/models/ae-real-nigi3d)
Model link: [View model](https://modelslab.com/models/ae-real-nigi3d)
View all models: [View Models](https://modelslab.com/models)
import requests
import json
url = "https://modelslab.com/api/v6/images/text2img"
payload = json.dumps({
"key": "your_api_key",
"model_id": "ae-real-nigi3d",
"prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
"negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
"width": "512",
"height": "512",
"samples": "1",
"num_inference_steps": "30",
"safety_checker": "no",
"enhance_prompt": "yes",
"seed": None,
"guidance_scale": 7.5,
"multi_lingual": "no",
"panorama": "no",
"self_attention": "no",
"upscale": "no",
"embeddings": "embeddings_model_id",
"lora": "lora_model_id",
"webhook": None,
"track_id": None
})
headers = {
'Content-Type': 'application/json'
}
response = requests.request("POST", url, headers=headers, data=payload)
print(response.text)
> Use this coupon code to get 25% off **DMGG0RBN** | {"license": "creativeml-openrail-m", "tags": ["modelslab.com", "stable-diffusion-api", "text-to-image", "ultra-realistic"], "pinned": true} | text-to-image | stablediffusionapi/ae-real-nigi3d | [
"diffusers",
"modelslab.com",
"stable-diffusion-api",
"text-to-image",
"ultra-realistic",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-06T04:40:10+00:00 | [] | [] | TAGS
#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
|
# ae-real-nigi3d API Inference
!generated from URL
## Get API Key
Get API key from ModelsLab API, No Payment needed.
Replace Key in below code, change model_id to "ae-real-nigi3d"
Coding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs
Try model for free: Generate Images
Model link: View model
View all models: View Models
import requests
import json
url = "URL
payload = URL({
"key": "your_api_key",
"model_id": "ae-real-nigi3d",
"prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
"negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
"width": "512",
"height": "512",
"samples": "1",
"num_inference_steps": "30",
"safety_checker": "no",
"enhance_prompt": "yes",
"seed": None,
"guidance_scale": 7.5,
"multi_lingual": "no",
"panorama": "no",
"self_attention": "no",
"upscale": "no",
"embeddings": "embeddings_model_id",
"lora": "lora_model_id",
"webhook": None,
"track_id": None
})
headers = {
'Content-Type': 'application/json'
}
response = requests.request("POST", url, headers=headers, data=payload)
print(URL)
> Use this coupon code to get 25% off DMGG0RBN | [
"# ae-real-nigi3d API Inference\n\n!generated from URL",
"## Get API Key\n\nGet API key from ModelsLab API, No Payment needed. \n\nReplace Key in below code, change model_id to \"ae-real-nigi3d\"\n\nCoding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs\n\nTry model for free: Generate Images\n\nModel link: View model\n\nView all models: View Models\n\n import requests \n import json \n \n url = \"URL \n \n payload = URL({ \n \"key\": \"your_api_key\", \n \"model_id\": \"ae-real-nigi3d\", \n \"prompt\": \"ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K\", \n \"negative_prompt\": \"painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime\", \n \"width\": \"512\", \n \"height\": \"512\", \n \"samples\": \"1\", \n \"num_inference_steps\": \"30\", \n \"safety_checker\": \"no\", \n \"enhance_prompt\": \"yes\", \n \"seed\": None, \n \"guidance_scale\": 7.5, \n \"multi_lingual\": \"no\", \n \"panorama\": \"no\", \n \"self_attention\": \"no\", \n \"upscale\": \"no\", \n \"embeddings\": \"embeddings_model_id\", \n \"lora\": \"lora_model_id\", \n \"webhook\": None, \n \"track_id\": None \n }) \n \n headers = { \n 'Content-Type': 'application/json' \n } \n \n response = requests.request(\"POST\", url, headers=headers, data=payload) \n \n print(URL)\n\n> Use this coupon code to get 25% off DMGG0RBN"
] | [
"TAGS\n#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n",
"# ae-real-nigi3d API Inference\n\n!generated from URL",
"## Get API Key\n\nGet API key from ModelsLab API, No Payment needed. \n\nReplace Key in below code, change model_id to \"ae-real-nigi3d\"\n\nCoding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs\n\nTry model for free: Generate Images\n\nModel link: View model\n\nView all models: View Models\n\n import requests \n import json \n \n url = \"URL \n \n payload = URL({ \n \"key\": \"your_api_key\", \n \"model_id\": \"ae-real-nigi3d\", \n \"prompt\": \"ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K\", \n \"negative_prompt\": \"painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime\", \n \"width\": \"512\", \n \"height\": \"512\", \n \"samples\": \"1\", \n \"num_inference_steps\": \"30\", \n \"safety_checker\": \"no\", \n \"enhance_prompt\": \"yes\", \n \"seed\": None, \n \"guidance_scale\": 7.5, \n \"multi_lingual\": \"no\", \n \"panorama\": \"no\", \n \"self_attention\": \"no\", \n \"upscale\": \"no\", \n \"embeddings\": \"embeddings_model_id\", \n \"lora\": \"lora_model_id\", \n \"webhook\": None, \n \"track_id\": None \n }) \n \n headers = { \n 'Content-Type': 'application/json' \n } \n \n response = requests.request(\"POST\", url, headers=headers, data=payload) \n \n print(URL)\n\n> Use this coupon code to get 25% off DMGG0RBN"
] | [
70,
19,
556
] | [
"passage: TAGS\n#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n# ae-real-nigi3d API Inference\n\n!generated from URL"
] | [
-0.052609775215387344,
0.052299387753009796,
-0.003810620168223977,
0.0510832741856575,
0.10892955213785172,
-0.03014417178928852,
0.18930283188819885,
0.012296602129936218,
0.023803889751434326,
0.03146379068493843,
0.13510791957378387,
0.1591504067182541,
0.0025721006095409393,
0.16411228477954865,
-0.07286085933446884,
-0.20985163748264313,
0.013605885207653046,
0.03745662420988083,
0.04770238324999809,
0.05327297002077103,
0.1273684799671173,
-0.020582616329193115,
0.13581903278827667,
0.0443725623190403,
-0.11483464390039444,
-0.011889960616827011,
-0.008361926302313805,
-0.07914038002490997,
0.046134114265441895,
0.08278923481702805,
-0.0021828364115208387,
0.0819544792175293,
0.011871772818267345,
-0.13690166175365448,
0.039366982877254486,
0.02088068798184395,
-0.06178973987698555,
0.04109682887792587,
-0.02831168845295906,
0.031918883323669434,
0.15630322694778442,
0.057007621973752975,
-0.0278113204985857,
0.03391212597489357,
-0.10645242035388947,
-0.03322581946849823,
0.025499071925878525,
0.03494685888290405,
0.08325603604316711,
0.01722399704158306,
0.05168171226978302,
0.04440116882324219,
0.028021445497870445,
0.0581362321972847,
0.09472217410802841,
-0.2619350254535675,
-0.027302242815494537,
0.20327110588550568,
0.08370191603899002,
0.004853839986026287,
-0.0454460047185421,
0.0848904699087143,
0.058541856706142426,
-0.019516441971063614,
0.05516448989510536,
-0.03079058602452278,
0.07500635087490082,
-0.0333702489733696,
-0.04648757353425026,
0.04811656475067139,
0.20524190366268158,
0.06593973934650421,
-0.019036728888750076,
-0.14616622030735016,
-0.09735693037509918,
0.1737736463546753,
-0.09282124787569046,
-0.00018810617621056736,
-0.0035397072788327932,
0.040912844240665436,
0.017333710566163063,
-0.08512354642152786,
-0.08228796720504761,
0.010405199602246284,
-0.0458044558763504,
0.13836874067783356,
-0.014790418557822704,
0.09842294454574585,
-0.08651965856552124,
0.09929550439119339,
-0.16000361740589142,
-0.13757842779159546,
-0.009690488688647747,
-0.16404181718826294,
0.08439073711633682,
0.05039922520518303,
0.02076880633831024,
-0.02947026863694191,
0.06547818332910538,
0.06984150409698486,
0.09400488436222076,
-0.021054644137620926,
0.09809892624616623,
0.1018134206533432,
0.027272937819361687,
-0.06926359236240387,
-0.10638613998889923,
-0.05657455325126648,
0.029169784858822823,
-0.013940142467617989,
0.058443762362003326,
-0.03524964302778244,
-0.12684175372123718,
-0.005143229383975267,
-0.1890401393175125,
-0.04145253449678421,
0.017363857477903366,
0.028000034391880035,
-0.08084210008382797,
-0.033425480127334595,
0.16652700304985046,
0.04018356278538704,
-0.03881127014756203,
-0.0795377716422081,
-0.023927636444568634,
0.22323602437973022,
0.12961247563362122,
0.018941165879368782,
0.021979782730340958,
0.06545424461364746,
-0.09023001044988632,
-0.030941275879740715,
-0.017988041043281555,
-0.027803869917988777,
0.013124693185091019,
-0.13130882382392883,
0.052280619740486145,
-0.12622591853141785,
-0.2016075700521469,
0.058608464896678925,
0.12397722899913788,
-0.053778037428855896,
-0.02877732925117016,
0.02605701982975006,
-0.03930328041315079,
0.04876355826854706,
0.004930430091917515,
-0.14321869611740112,
-0.05396208167076111,
0.015203538350760937,
0.0013538080966100097,
0.11418449133634567,
-0.2062682956457138,
0.035406485199928284,
-0.02180526591837406,
0.02175886370241642,
-0.07592485845088959,
0.03242690488696098,
-0.031735680997371674,
0.05120690166950226,
-0.05620061233639717,
-0.037299223244190216,
-0.003251644317060709,
0.022369345650076866,
0.05851252004504204,
0.1808566153049469,
-0.0632716566324234,
-0.0048993597738444805,
0.09740854799747467,
-0.08302915096282959,
-0.15263471007347107,
0.0609542541205883,
-0.007150334771722555,
0.11168055236339569,
0.037391237914562225,
0.020177707076072693,
-0.007699442096054554,
-0.28388282656669617,
0.11910304427146912,
0.10199551284313202,
-0.053593266755342484,
-0.13112448155879974,
0.031559087336063385,
0.07457810640335083,
0.019529150798916817,
0.06527639180421829,
-0.05435878410935402,
0.10844369977712631,
-0.07966218888759613,
-0.026992900297045708,
-0.03322749584913254,
-0.06808656454086304,
0.008737591095268726,
0.03677893802523613,
0.021837634965777397,
-0.0008344585658051074,
-0.03363143652677536,
-0.006888733245432377,
0.04367511346936226,
0.026109812781214714,
-0.027731293812394142,
-0.05649017542600632,
0.15405447781085968,
-0.14475923776626587,
-0.014764796011149883,
-0.04053923860192299,
-0.0411338247358799,
-0.01966634951531887,
0.10135391354560852,
0.029394039884209633,
0.13076171278953552,
0.07489630579948425,
0.07245621830224991,
-0.001159036299213767,
-0.01718478836119175,
0.03694652393460274,
0.03549109026789665,
-0.020597701892256737,
-0.16183756291866302,
0.07068677246570587,
-0.08529402315616608,
-0.010086288675665855,
-0.18253636360168457,
0.014929627999663353,
-0.07555145025253296,
0.11857452988624573,
0.07053299248218536,
-0.0011065291473641992,
0.004638846032321453,
-0.05652385205030441,
-0.08577707409858704,
-0.0039819697849452496,
0.07465619593858719,
0.03595403954386711,
-0.012164178304374218,
0.19808459281921387,
-0.03219103440642357,
0.30539172887802124,
0.10653943568468094,
-0.10218609869480133,
-0.02504153363406658,
-0.06845346838235855,
-0.04867742210626602,
0.037944432348012924,
-0.007156759034842253,
0.021620335057377815,
-0.021542159840464592,
-0.005256361328065395,
0.13237355649471283,
-0.06741973012685776,
0.038104552775621414,
0.08507419377565384,
-0.1096920296549797,
-0.06472698599100113,
0.05504953861236572,
0.154364213347435,
-0.10632278025150299,
0.10350484400987625,
0.17827872931957245,
0.005846597254276276,
0.15044786036014557,
0.04410465806722641,
-0.045895423740148544,
-0.030628172680735588,
0.05749741569161415,
-0.0007496252073906362,
0.19292154908180237,
-0.046754609793424606,
0.017935339361429214,
0.06436818093061447,
-0.04538382217288017,
0.026563074439764023,
-0.07474509626626968,
-0.06788339465856552,
0.061577506363391876,
0.007619043346494436,
0.09221586585044861,
0.12855270504951477,
-0.12619009613990784,
0.11461053043603897,
-0.052189894020557404,
-0.07520875334739685,
0.036453746259212494,
0.00013121642405167222,
-0.029163511469960213,
0.09552283585071564,
-0.029173394665122032,
-0.09211455285549164,
-0.15716922283172607,
-0.20343594253063202,
-0.10718420147895813,
-0.0016998822102323174,
0.03122834488749504,
-0.031643204391002655,
-0.07503794878721237,
-0.05495378002524376,
-0.06246121600270271,
0.009906979277729988,
-0.03406522050499916,
0.01408535148948431,
-0.016431625932455063,
-0.05267427861690521,
-0.07636648416519165,
-0.04135815426707268,
-0.029209095984697342,
0.11253781616687775,
0.08598112314939499,
-0.057676926255226135,
0.152284637093544,
0.09048058092594147,
0.0018123375484719872,
0.019459787756204605,
0.07414624094963074,
0.2518569231033325,
-0.005129607394337654,
0.12911304831504822,
0.28807005286216736,
0.08672472089529037,
0.10709239542484283,
0.11720769107341766,
0.05609872192144394,
-0.08213818818330765,
0.025310982018709183,
-0.06259207427501678,
-0.11946846544742584,
-0.08717509359121323,
-0.11747679859399796,
-0.10751118510961533,
-0.061671070754528046,
0.0034558901097625494,
0.022188017144799232,
0.08079775422811508,
0.17186211049556732,
0.021227674558758736,
0.05315077677369118,
0.021095335483551025,
0.09238964319229126,
0.09599088132381439,
-0.027025775983929634,
0.08364813774824142,
-0.0799214169383049,
-0.050535283982753754,
0.12884056568145752,
0.05006859451532364,
0.04071200639009476,
-0.025820253416895866,
-0.02448332868516445,
0.10623405873775482,
0.020030731335282326,
0.10961543768644333,
0.09605975449085236,
-0.05658555403351784,
-0.05243079364299774,
-0.044517338275909424,
-0.07892140001058578,
0.060081034898757935,
0.11480522155761719,
-0.06323476880788803,
-0.06906381994485855,
0.028594356030225754,
0.05583667382597923,
0.03026644140481949,
0.03810930624604225,
0.16940446197986603,
-0.3529067039489746,
-0.0073175979778170586,
0.021031195297837257,
0.09968508780002594,
-0.08229396492242813,
0.005644495598971844,
0.11625760793685913,
-0.039090048521757126,
0.08028542995452881,
-0.05305337905883789,
0.09832195937633514,
0.04886901378631592,
-0.010144402272999287,
0.0677739828824997,
0.0862518921494484,
0.01519935205578804,
0.02704060636460781,
-0.15732090175151825,
0.12267249077558517,
-0.004940046928822994,
0.045409731566905975,
-0.028188781812787056,
-0.013247680850327015,
0.031486980617046356,
0.12732866406440735,
0.188742995262146,
0.007785203866660595,
0.17483271658420563,
0.015642596408724785,
-0.08278772979974747,
-0.011656597256660461,
0.0388491228222847,
0.02644575573503971,
-0.013512403704226017,
0.041668541729450226,
-0.011538996361196041,
0.020125091075897217,
0.059952378273010254,
-0.18102459609508514,
-0.17768345773220062,
0.034069690853357315,
0.11596240103244781,
-0.06904131919145584,
-0.007512206677347422,
0.004182620905339718,
-0.052954334765672684,
0.19016553461551666,
0.013677021488547325,
-0.11292725056409836,
-0.14663270115852356,
-0.09909869730472565,
0.00550440326333046,
-0.043719582259655,
0.06099724397063255,
-0.12881304323673248,
0.043491046875715256,
-0.05656735599040985,
-0.14673814177513123,
0.08446478843688965,
-0.14079973101615906,
0.005542283412069082,
-0.12086423486471176,
0.020835772156715393,
-0.02008376084268093,
-0.07537157088518143,
0.01887303963303566,
-0.04769669845700264,
-0.06682085990905762,
-0.13741590082645416,
0.020686734467744827,
0.1284468024969101,
-0.046256519854068756,
0.011162438429892063,
-0.09157133847475052,
-0.02057684399187565,
0.03623025119304657,
0.03936077281832695,
0.059957221150398254,
0.20054477453231812,
-0.07665106654167175,
0.06359051167964935,
0.2373906970024109,
-0.05073212459683418,
-0.2124546766281128,
-0.08650436252355576,
-0.10076936334371567,
0.005099095404148102,
-0.013957267627120018,
-0.07001364976167679,
0.1436574012041092,
-0.012103566899895668,
-0.038051970303058624,
0.2017240971326828,
-0.29356321692466736,
-0.08309569209814072,
0.02504364587366581,
0.136499285697937,
0.3248675763607025,
-0.12903131544589996,
-0.03731001913547516,
-0.057215530425310135,
-0.30479636788368225,
0.0898551419377327,
-0.002869906835258007,
0.05028438940644264,
-0.07926779240369797,
0.042813412845134735,
-0.02049042098224163,
-0.06247115507721901,
0.1438651978969574,
-0.04798929765820503,
0.0751621425151825,
-0.1207231730222702,
0.11867568641901016,
0.16008664667606354,
-0.03565683215856552,
0.11090172082185745,
-0.13539375364780426,
0.09696540236473083,
-0.1389746069908142,
-0.047090474516153336,
-0.047575660049915314,
0.03591063991189003,
-0.012901870533823967,
-0.12663666903972626,
-0.07151178270578384,
0.004374935757368803,
0.04348970204591751,
0.028826892375946045,
-0.012268445454537868,
0.004618629813194275,
0.04543953388929367,
0.15183842182159424,
-0.03669648617506027,
-0.08139272034168243,
-0.13771376013755798,
-0.07917959988117218,
-0.026108641177415848,
0.08607156574726105,
-0.18992415070533752,
-0.051596835255622864,
0.14964641630649567,
0.034892115741968155,
0.06214829534292221,
0.046040233224630356,
-0.03387896344065666,
0.06492209434509277,
0.09866023063659668,
-0.14512260258197784,
-0.023617295548319817,
-0.050035394728183746,
0.19240769743919373,
0.11454372853040695,
0.0960150882601738,
0.11875244975090027,
-0.07925907522439957,
0.05294010788202286,
-0.02119370363652706,
0.024894429370760918,
-0.04564332962036133,
0.021612608805298805,
0.037205327302217484,
-0.0027798961382359266,
-0.05895516276359558,
0.029500259086489677,
-0.07724562287330627,
-0.08167348057031631,
-0.09445887058973312,
-0.03251349925994873,
-0.10486970096826553,
-0.0612315833568573,
0.055626243352890015,
0.04321874678134918,
-0.2070913165807724,
-0.016564933583140373,
0.012054212391376495,
-0.12061426043510437,
0.04218351095914841,
0.03951450437307358,
0.047003813087940216,
-0.021985838189721107,
-0.02367229573428631,
-0.07308991253376007,
-0.04065860062837601,
0.0016070741694420576,
0.056128792464733124,
0.09955734014511108,
-0.1883193850517273,
-0.23467408120632172,
-0.04562995582818985,
0.03458453342318535,
-0.08799611032009125,
-0.023129217326641083,
-0.09489551931619644,
-0.005476740654557943,
-0.1410244256258011,
0.02261173538863659,
-0.13916604220867157,
-0.039290349930524826,
-0.04401953145861626,
-0.05624621734023094,
-0.028878306970000267,
0.029879242181777954,
-0.05239569768309593,
0.011638536117970943,
0.025515275076031685,
0.02869703248143196,
-0.05129917711019516,
-0.02432025969028473,
-0.01317758671939373,
-0.09249746799468994,
0.07759685814380646,
-0.036086153239011765,
-0.11930517107248306,
-0.08872810751199722,
-0.22250010073184967,
-0.018048668280243874,
0.09429008513689041,
0.018893936648964882,
0.016041524708271027,
0.09315261244773865,
0.08871152997016907,
0.02489374205470085,
-0.01036971714347601,
-0.06505589932203293,
0.06191125512123108,
-0.07925883680582047,
-0.016565747559070587,
-0.09215553104877472,
0.007998188026249409,
-0.07338815927505493,
0.006184850819408894,
0.1632571518421173,
0.07109387964010239,
0.14979900419712067,
-0.05075404420495033,
0.040890879929065704,
-0.06892168521881104,
0.015161576680839062,
0.057056766003370285,
-0.04537549614906311,
0.08874650299549103,
-0.042010337114334106,
-0.053796373307704926,
-0.03824171796441078,
0.24802665412425995,
-0.021612588316202164,
-0.17901776731014252,
-0.0022778790444135666,
-0.0329272523522377,
0.011912512592971325,
-0.022144710645079613,
0.15368743240833282,
0.00924486480653286,
0.05234602466225624,
-0.19460682570934296,
0.06682806462049484,
0.0989769771695137,
0.021001366898417473,
-0.026396658271551132,
0.16048473119735718,
-0.041381463408470154,
0.11488794535398483,
0.04122380167245865,
0.022728370502591133,
-0.03045712225139141,
0.003481414634734392,
-0.057222798466682434,
0.13894711434841156,
-0.015941888093948364,
-0.03802099451422691,
0.17464977502822876,
-0.013847963884472847,
-0.00469904113560915,
0.07092171162366867,
-0.03244587406516075,
-0.05637270584702492,
-0.2144032120704651,
-0.051203224807977676,
-0.13666802644729614,
0.03981653228402138,
-0.04051361605525017,
0.026394326239824295,
0.022511569783091545,
0.0697745680809021,
-0.026536762714385986,
-0.013413816690444946,
-0.05274783819913864,
-0.0618552602827549,
0.13101916015148163,
-0.0198429636657238,
-0.06852564215660095,
0.04233843460679054,
0.019199209287762642,
-0.07745096832513809,
-0.030963364988565445,
-0.03770962730050087,
0.07931269705295563,
0.02624637261033058,
0.051233045756816864,
0.005794217344373465,
-0.03360286355018616,
-0.03780461475253105,
0.022497907280921936,
-0.016128061339259148,
0.12502513825893402,
-0.01192561537027359,
-0.00023443633108399808,
0.0053444961085915565,
0.07787514477968216,
-0.019326699897646904,
-0.042056482285261154,
-0.11182232201099396,
0.07456687092781067,
-0.0871167853474617,
0.09681565314531326,
-0.07156208157539368,
-0.00917016714811325,
-0.028504563495516777,
0.24848316609859467,
0.20281071960926056,
-0.191972553730011,
-0.00775787653401494,
-0.03745449334383011,
0.00438105221837759,
-0.01016288809478283,
0.05473167821764946,
0.014138002879917622,
0.3229914903640747,
-0.037413328886032104,
-0.04081055149435997,
-0.1087481826543808,
-0.012007522396743298,
-0.026900840923190117,
-0.03146766871213913,
0.026695402339100838,
-0.07857313752174377,
-0.09420116990804672,
0.07909499108791351,
-0.19331829249858856,
0.05004028603434563,
0.118763767182827,
-0.05159413814544678,
-0.016360152512788773,
-0.05751394107937813,
-0.03130752220749855,
0.004335664212703705,
0.05435744300484657,
-0.10979990661144257,
-0.043233733624219894,
0.0091581204906106,
-0.00011036424984922633,
-0.10506316274404526,
0.08688283711671829,
0.0009994166903197765,
-0.01390941720455885,
0.08220546692609787,
0.00586279621347785,
0.002933025360107422,
0.0755237340927124,
-0.017710424959659576,
-0.07388632744550705,
0.08189617842435837,
-0.010404872708022594,
-0.029866864904761314,
-0.03636656701564789,
-0.003865605453029275,
-0.00020867265993729234,
-0.028225665912032127,
0.007666412275284529,
-0.13140127062797546,
0.013722089119255543,
0.0748923122882843,
-0.08962897956371307,
-0.07730364799499512,
0.047708604484796524,
-0.0548650324344635,
0.08713427931070328,
0.001330494531430304,
-0.011966222897171974,
-0.04794875159859657,
-0.0031088110990822315,
0.11352677643299103,
0.011857427656650543,
-0.1346464604139328,
0.028129789978265762,
-0.040730156004428864,
-0.026766663417220116,
-0.056402768939733505,
0.058438822627067566,
-0.150656059384346,
-0.0209307000041008,
-0.0996365174651146,
0.027655893936753273,
-0.030724884942173958,
0.08069554716348648,
0.2848212718963623,
0.013101794756948948,
-0.025768334046006203,
-0.10309739410877228,
0.009459150023758411,
0.05352596566081047,
0.012311723083257675,
-0.09506849199533463
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my_article_model2
This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.8307
- Rouge1: 0.002
- Rouge2: 0.0011
- Rougel: 0.002
- Rougelsum: 0.002
- Gen Len: 0.095
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
| No log | 1.0 | 50 | 2.2217 | 0.0401 | 0.017 | 0.0369 | 0.0372 | 2.09 |
| No log | 2.0 | 100 | 1.8971 | 0.001 | 0.0005 | 0.001 | 0.001 | 0.095 |
| No log | 3.0 | 150 | 1.8432 | 0.001 | 0.0005 | 0.001 | 0.001 | 0.095 |
| No log | 4.0 | 200 | 1.8307 | 0.002 | 0.0011 | 0.002 | 0.002 | 0.095 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["rouge"], "base_model": "t5-small", "model-index": [{"name": "my_article_model2", "results": []}]} | text2text-generation | hussainBurhan/my_article_model2 | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:t5-small",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T04:42:14+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| my\_article\_model2
===================
This model is a fine-tuned version of t5-small on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 1.8307
* Rouge1: 0.002
* Rouge2: 0.0011
* Rougel: 0.002
* Rougelsum: 0.002
* Gen Len: 0.095
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 4
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 4",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 4",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
77,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 4### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09460937976837158,
0.06795691698789597,
-0.0019046663073822856,
0.10757317394018173,
0.12441359460353851,
-0.0025696696247905493,
0.16910600662231445,
0.12435691803693771,
-0.06783522665500641,
0.0454556979238987,
0.1358487606048584,
0.11368320882320404,
0.031879741698503494,
0.14064987003803253,
-0.06671234220266342,
-0.20283597707748413,
0.013972509652376175,
0.025985607877373695,
-0.048011619597673416,
0.1282389760017395,
0.0975501760840416,
-0.11177538335323334,
0.10798180103302002,
-0.011087257415056229,
-0.1532382220029831,
0.0022942887153476477,
0.018311727792024612,
-0.052219025790691376,
0.13803791999816895,
0.04287729039788246,
0.09619583934545517,
0.035016778856515884,
0.06319307535886765,
-0.17537233233451843,
0.012037922628223896,
0.06269032508134842,
-0.0045380056835711,
0.08487307280302048,
0.05178317427635193,
0.014781962148845196,
0.09258580207824707,
-0.08502243459224701,
0.04331914335489273,
0.031987812370061874,
-0.12455437332391739,
-0.198011577129364,
-0.0750550702214241,
0.032399918884038925,
0.09957511723041534,
0.0949028879404068,
-0.017219454050064087,
0.13583773374557495,
-0.02224981039762497,
0.10345850884914398,
0.2130099982023239,
-0.3191310465335846,
-0.06266102194786072,
0.0459858663380146,
0.06220631301403046,
0.11650863289833069,
-0.08977103233337402,
-0.009625478647649288,
0.05691077187657356,
0.03271043673157692,
0.14984187483787537,
-0.02739824540913105,
-0.027436600998044014,
-0.004837059881538153,
-0.13982780277729034,
-0.036652740091085434,
0.17895765602588654,
0.058055732399225235,
-0.04748921096324921,
-0.07313140481710434,
-0.0821676254272461,
-0.13957808911800385,
-0.0327611044049263,
-0.015699947252869606,
0.04538658633828163,
-0.009964426048099995,
-0.05933850258588791,
-0.04590042680501938,
-0.11530444025993347,
-0.0654374435544014,
-0.033364202827215195,
0.08656765520572662,
0.0236021988093853,
-0.003010132582858205,
-0.03695075213909149,
0.0886954516172409,
-0.018304092809557915,
-0.13918063044548035,
0.015460895374417305,
0.021573450416326523,
0.02421957440674305,
-0.038795433938503265,
-0.04828460142016411,
-0.10772908478975296,
0.027148908004164696,
0.11622920632362366,
-0.0452176108956337,
0.049240950495004654,
-0.003090529702603817,
0.04729459434747696,
-0.1101933941245079,
0.15424709022045135,
-0.03515876829624176,
-0.05532927066087723,
0.03148967772722244,
0.09803950786590576,
0.06646070629358292,
-0.015585631132125854,
-0.139017716050148,
0.024171002209186554,
0.11965727806091309,
0.028544843196868896,
-0.03620976209640503,
0.06682313978672028,
-0.041558749973773956,
-0.011061611585319042,
0.031299758702516556,
-0.0833505392074585,
0.0033391909673810005,
-0.011506634764373302,
-0.04347575828433037,
-0.06415054947137833,
0.026456959545612335,
0.019625931978225708,
-0.011830458417534828,
0.07193683832883835,
-0.08740285038948059,
-0.008469022810459137,
-0.06949257850646973,
-0.12208230793476105,
0.020408766344189644,
-0.07613063603639603,
0.017078466713428497,
-0.11922883242368698,
-0.19345562160015106,
-0.0019863832276314497,
0.05034128576517105,
-0.032147154211997986,
-0.05512594059109688,
-0.05922137573361397,
-0.0861770436167717,
0.02164984494447708,
-0.01778634823858738,
0.09547431766986847,
-0.07025865465402603,
0.09921079874038696,
0.04116084799170494,
0.05916473641991615,
-0.057202745229005814,
0.03268447890877724,
-0.10731402039527893,
0.037692636251449585,
-0.15499743819236755,
0.02715272083878517,
-0.03054356388747692,
0.07825280725955963,
-0.10086922347545624,
-0.06725240498781204,
-0.012654559686779976,
-0.001793450559489429,
0.07163872569799423,
0.10416526347398758,
-0.1642179787158966,
-0.05763474479317665,
0.18017463386058807,
-0.0871061161160469,
-0.17880910634994507,
0.14516068994998932,
-0.0357804149389267,
0.06904919445514679,
0.06831423938274384,
0.20566464960575104,
0.0570216029882431,
-0.08806266635656357,
0.009385615587234497,
0.007849385030567646,
0.06789372861385345,
-0.0643053874373436,
0.09223206341266632,
-0.007595490198582411,
0.0029229437932372093,
0.005207831505686045,
-0.04400018975138664,
0.05887625738978386,
-0.065128855407238,
-0.07782551646232605,
-0.04220752790570259,
-0.10658404976129532,
0.036032866686582565,
0.03957224264740944,
0.05924179404973984,
-0.12265637516975403,
-0.09652595967054367,
0.03463387116789818,
0.06941022723913193,
-0.07742703706026077,
0.02262035943567753,
-0.07201681286096573,
0.09402276575565338,
-0.08773994445800781,
-0.009059472009539604,
-0.1380477100610733,
-0.04253539443016052,
0.01730601117014885,
0.010734480805695057,
0.012389467097818851,
0.0012724383268505335,
0.0788780078291893,
0.07724791020154953,
-0.06766195595264435,
-0.03345157951116562,
-0.008199078030884266,
0.013321731239557266,
-0.12557679414749146,
-0.18856006860733032,
-0.017182564362883568,
-0.02714073285460472,
0.14964787662029266,
-0.23119224607944489,
0.0537225715816021,
0.005301622208207846,
0.06724535673856735,
0.03812275826931,
-0.012006428092718124,
-0.025945832952857018,
0.04692695662379265,
-0.05159183219075203,
-0.07241853326559067,
0.07087220251560211,
0.027998005971312523,
-0.10621678829193115,
0.0015880075516179204,
-0.1798967719078064,
0.18336287140846252,
0.133807972073555,
-0.05661654844880104,
-0.06298328191041946,
0.011201740242540836,
-0.03686174750328064,
-0.03192668408155441,
-0.029929613694548607,
-0.023658322170376778,
0.1040562093257904,
-0.00027027781470678747,
0.16567206382751465,
-0.10486030578613281,
-0.048864610493183136,
0.03229282796382904,
-0.04318274185061455,
0.013845192268490791,
0.11649985611438751,
0.06375942379236221,
-0.1360902190208435,
0.15429364144802094,
0.19800102710723877,
-0.05931559205055237,
0.14141729474067688,
-0.04621186479926109,
-0.06738005578517914,
-0.022787844762206078,
0.032350435853004456,
0.023754822090268135,
0.10480162501335144,
-0.10345781594514847,
0.008720320649445057,
0.014013255015015602,
0.020484432578086853,
0.014273540116846561,
-0.20257295668125153,
-0.024586087092757225,
0.05313654616475105,
-0.05635667219758034,
0.0005316378665156662,
-0.01190427877008915,
-0.02291269786655903,
0.09857354313135147,
0.004786136094480753,
-0.07144305109977722,
0.04422484710812569,
0.0003645613614935428,
-0.08850474655628204,
0.20447438955307007,
-0.06664673238992691,
-0.16814589500427246,
-0.14514297246932983,
-0.06270246207714081,
-0.053081072866916656,
0.03114607371389866,
0.07751879841089249,
-0.06282788515090942,
-0.03897775709629059,
-0.13388468325138092,
0.007588637992739677,
0.010538442060351372,
0.018914995715022087,
0.012485787272453308,
-0.0029147895984351635,
0.0837981179356575,
-0.1023692637681961,
-0.011091538704931736,
-0.01274728961288929,
-0.03891490399837494,
0.028159823268651962,
0.006139382254332304,
0.11410345137119293,
0.1354726403951645,
-0.013944164849817753,
-0.004120022989809513,
-0.039396267384290695,
0.23577263951301575,
-0.060894083231687546,
0.006481500342488289,
0.1578073650598526,
-0.016167178750038147,
0.05706721544265747,
0.13038691878318787,
0.04142582044005394,
-0.09957928210496902,
0.025107325986027718,
0.022057177498936653,
-0.03422604873776436,
-0.21191923320293427,
-0.01663336157798767,
-0.050500210374593735,
0.03400672599673271,
0.09377260506153107,
0.03335648030042648,
0.04797127842903137,
0.06544725596904755,
0.010288732126355171,
0.08891982585191727,
0.003557156538590789,
0.08390142023563385,
0.11304455995559692,
0.045848552137613297,
0.1290193498134613,
-0.05921623483300209,
-0.038341157138347626,
0.0458596833050251,
0.008948490023612976,
0.1795874387025833,
0.014150450937449932,
0.2002842128276825,
0.04464474692940712,
0.14462833106517792,
-0.00333011313341558,
0.07587087899446487,
-0.016619792208075523,
-0.0326450876891613,
-0.023600684478878975,
-0.05565355345606804,
-0.028291255235671997,
0.030358023941516876,
-0.09533712267875671,
0.058252349495887756,
-0.1046615019440651,
0.024874137714505196,
0.049895431846380234,
0.2536342442035675,
0.057525746524333954,
-0.3426858186721802,
-0.10482021421194077,
0.02741485834121704,
-0.02388937585055828,
-0.03516068682074547,
0.026682117953896523,
0.15338623523712158,
-0.05395173653960228,
0.03372366353869438,
-0.0810522511601448,
0.08999887853860855,
-0.026378370821475983,
0.05224328115582466,
0.045266877859830856,
0.0744393989443779,
-0.012537011876702309,
0.071112722158432,
-0.28318631649017334,
0.2631070911884308,
0.017246240749955177,
0.06587830185890198,
-0.056586477905511856,
0.002014587400481105,
0.02245595119893551,
0.07095903158187866,
0.08756615221500397,
-0.015333653427660465,
-0.05967766046524048,
-0.1724015772342682,
-0.07784071564674377,
0.025754868984222412,
0.08744317293167114,
-0.04911188781261444,
0.10896381735801697,
-0.05141250044107437,
-0.0006344548310153186,
0.08038783073425293,
0.006516889203339815,
-0.06174783036112785,
-0.10266958922147751,
-0.004036939702928066,
0.05066405609250069,
-0.02946949377655983,
-0.08086851239204407,
-0.0848340392112732,
-0.09893426299095154,
0.15093185007572174,
-0.04745050147175789,
-0.04724593088030815,
-0.10072507709264755,
0.039416998624801636,
0.047600459307432175,
-0.08257127553224564,
0.0426512211561203,
0.004129324108362198,
0.09690394997596741,
0.016210757195949554,
-0.06692945957183838,
0.12750273942947388,
-0.07821810990571976,
-0.1737251579761505,
-0.05430645868182182,
0.1268128603696823,
-0.008431154303252697,
0.0443989522755146,
-0.0014089415781199932,
0.009092790074646473,
-0.034679483622312546,
-0.0658278614282608,
0.016753334552049637,
-0.015932809561491013,
0.05492977052927017,
-0.01995060220360756,
-0.03179715946316719,
0.01248831395059824,
-0.058467742055654526,
-0.05088427662849426,
0.16980181634426117,
0.2991796135902405,
-0.07282992452383041,
0.010335809551179409,
0.05647352710366249,
-0.05974796786904335,
-0.18216407299041748,
0.01627984642982483,
0.015978137031197548,
0.002504331525415182,
0.0629284456372261,
-0.1274777203798294,
0.07515998184680939,
0.084409698843956,
-0.025453953072428703,
0.10847653448581696,
-0.32043179869651794,
-0.13797062635421753,
0.09865059703588486,
0.15764816105365753,
0.12287692725658417,
-0.17132623493671417,
-0.04796022176742554,
-0.043378766626119614,
-0.12140870839357376,
0.11341075599193573,
-0.14018361270427704,
0.11078903824090958,
-0.005114836618304253,
0.049333252012729645,
0.004899098537862301,
-0.05052272230386734,
0.12454899400472641,
-0.039851363748311996,
0.09174540638923645,
-0.0743555799126625,
0.01781744323670864,
0.07354952394962311,
-0.06524767726659775,
0.02889862097799778,
-0.155457004904747,
0.03757327049970627,
-0.06527691334486008,
-0.028868015855550766,
-0.053250569850206375,
0.027468331158161163,
-0.038450904190540314,
-0.06271544098854065,
-0.03496512025594711,
0.015090955421328545,
0.05395675450563431,
-0.012525459751486778,
0.1530231535434723,
0.0011111695785075426,
0.15157197415828705,
0.15289345383644104,
0.0943569540977478,
-0.06841173768043518,
-0.021171556785702705,
-0.009304218925535679,
-0.03885180130600929,
0.04856856167316437,
-0.1639387607574463,
0.03235926106572151,
0.10862789303064346,
0.00015567905211355537,
0.13796181976795197,
0.0665789470076561,
-0.03556133061647415,
0.011120500974357128,
0.06512627750635147,
-0.17080678045749664,
-0.14636167883872986,
-0.030860766768455505,
-0.0021072104573249817,
-0.13010671734809875,
0.045720670372247696,
0.13390998542308807,
-0.07686673849821091,
0.006614017300307751,
-0.00910166185349226,
0.021214686334133148,
-0.03408092260360718,
0.16403259336948395,
0.04777234420180321,
0.055485162883996964,
-0.07444987446069717,
0.08650495111942291,
0.05438665300607681,
-0.07936795055866241,
0.024089345708489418,
0.06331261992454529,
-0.09137335419654846,
-0.04953349754214287,
0.05549250543117523,
0.16686055064201355,
-0.033004358410835266,
-0.05373721197247505,
-0.14828692376613617,
-0.11538790911436081,
0.05169909447431564,
0.1739545613527298,
0.07930216938257217,
0.01384124718606472,
-0.023429781198501587,
0.005023417063057423,
-0.11123853921890259,
0.12483750283718109,
0.03402628004550934,
0.08516191691160202,
-0.15803004801273346,
0.1118212640285492,
-0.008519550785422325,
0.013275707140564919,
-0.018902020528912544,
0.04961397126317024,
-0.10049670934677124,
-0.007024445105344057,
-0.13676244020462036,
0.0049004401080310345,
-0.022178594022989273,
-0.0036961750593036413,
-0.010122470557689667,
-0.04469874128699303,
-0.06183462217450142,
0.015095800161361694,
-0.09593931585550308,
-0.03322932496666908,
0.027478009462356567,
0.055686090141534805,
-0.11603982746601105,
-0.03315626084804535,
0.022378526628017426,
-0.0771339014172554,
0.07377991825342178,
0.006146571133285761,
0.008419996127486229,
0.04583141580224037,
-0.1464008241891861,
0.03890408203005791,
0.049799468368291855,
0.011808368377387524,
0.03593941032886505,
-0.08027005940675735,
-0.02240118943154812,
0.0025125963147729635,
0.03550570830702782,
0.017945149913430214,
0.09376801550388336,
-0.12023758888244629,
0.004932340234518051,
-0.017176007851958275,
-0.04368813335895538,
-0.05723594129085541,
0.02931881695985794,
0.06306266784667969,
0.00776277668774128,
0.20621830224990845,
-0.09289983659982681,
0.012247364036738873,
-0.21024064719676971,
0.009274483658373356,
0.00462682917714119,
-0.125503808259964,
-0.11390524357557297,
-0.05970250070095062,
0.04805172234773636,
-0.062351979315280914,
0.10643962770700455,
-0.005691525526344776,
0.05148295685648918,
0.03925232216715813,
-0.02440117672085762,
0.05289299413561821,
0.017258265987038612,
0.24153092503547668,
0.007953467778861523,
-0.04157340154051781,
0.024719053879380226,
0.026403242722153664,
0.11139024794101715,
0.09191595017910004,
0.1814737617969513,
0.1627684384584427,
-0.04583735018968582,
0.11051762104034424,
0.029714444652199745,
-0.032638609409332275,
-0.15623392164707184,
0.03780096396803856,
-0.02832438051700592,
0.11599528789520264,
-0.02093503810465336,
0.2241109013557434,
0.12530554831027985,
-0.15233972668647766,
0.008954537101089954,
-0.06652908772230148,
-0.06887812167406082,
-0.10411636531352997,
-0.08914049714803696,
-0.09314878284931183,
-0.13843567669391632,
-0.02041785791516304,
-0.1094285249710083,
0.0184348002076149,
0.10372980684041977,
0.011504980735480785,
-0.027020303532481194,
0.1758149415254593,
0.012551650404930115,
0.008328159339725971,
0.048509079962968826,
-0.004162218421697617,
-0.03271748498082161,
-0.07765725255012512,
-0.09203120321035385,
0.011267120949923992,
-0.010893464088439941,
0.02681685797870159,
-0.038652416318655014,
-0.026217615231871605,
0.041561439633369446,
-0.019220763817429543,
-0.11139651387929916,
0.008725344203412533,
0.03393389284610748,
0.051712967455387115,
0.051526304334402084,
0.013700085692107677,
-0.0041341292671859264,
0.004068394191563129,
0.24330513179302216,
-0.08272870630025864,
-0.07672558724880219,
-0.08272935450077057,
0.19029325246810913,
0.00631882855668664,
-0.007931825704872608,
0.003744421061128378,
-0.08784306794404984,
0.03001716360449791,
0.2360251396894455,
0.17662183940410614,
-0.08749588578939438,
-0.0059721218422055244,
-0.021092304959893227,
-0.006011325400322676,
-0.037035636603832245,
0.1053384318947792,
0.11002957075834274,
0.017983324825763702,
-0.07940201461315155,
-0.01974259316921234,
-0.035967838019132614,
0.001446202164515853,
-0.04935723915696144,
0.07544953376054764,
0.020553775131702423,
0.009456089697778225,
-0.028289133682847023,
0.06679190695285797,
-0.03304927423596382,
-0.09377548843622208,
-0.001516711781732738,
-0.1921107918024063,
-0.1383901685476303,
-0.026450524106621742,
0.11442851275205612,
-0.017450744286179543,
0.04527674615383148,
-0.029006212949752808,
0.02113799750804901,
0.045481182634830475,
-0.02336212433874607,
-0.06787750124931335,
-0.05757122114300728,
0.06104698404669762,
-0.12474331259727478,
0.22252948582172394,
-0.0375346764922142,
0.024778807535767555,
0.13298457860946655,
0.03210173174738884,
-0.09887655079364777,
0.09426998347043991,
0.051034606993198395,
-0.03180674463510513,
0.05021987110376358,
0.08660230785608292,
-0.03332606703042984,
0.11155348271131516,
0.05191536992788315,
-0.12650202214717865,
0.005959984380751848,
-0.03557787835597992,
-0.055703166872262955,
-0.040067825466394424,
-0.05078265070915222,
-0.057920970022678375,
0.1398380696773529,
0.16612273454666138,
-0.05628111958503723,
-0.002365127671509981,
-0.05234720930457115,
0.01557513140141964,
0.07675217092037201,
0.03393397852778435,
-0.023416878655552864,
-0.23301291465759277,
0.0054295863956213,
0.07291677594184875,
0.004388659726828337,
-0.31409379839897156,
-0.08006133884191513,
-0.018899204209446907,
-0.03769034147262573,
-0.11424104869365692,
0.08881005644798279,
0.13537214696407318,
0.04012222960591316,
-0.06251368671655655,
-0.054869506508111954,
-0.07862477004528046,
0.1660228818655014,
-0.12396766990423203,
-0.09746246784925461
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | dvilasuero/DistilabelOpenHermes-2.5-mistral-7b-mix2 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T04:42:30+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
60,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04571164771914482,
0.1637648642063141,
-0.005522117950022221,
0.017756497487425804,
0.09821303188800812,
0.01318030059337616,
0.06541220843791962,
0.1127115860581398,
-0.017605241388082504,
0.1127321794629097,
0.030432263389229774,
0.09820804744958878,
0.1134178638458252,
0.14702944457530975,
-0.003594378475099802,
-0.22472713887691498,
0.052083637565374374,
-0.12124937027692795,
-0.03241228312253952,
0.1181139275431633,
0.14941681921482086,
-0.09871039539575577,
0.07234785705804825,
-0.030714161694049835,
-0.01334790326654911,
-0.03167412802577019,
-0.05947697162628174,
-0.045681875199079514,
0.046136777848005295,
0.0657167062163353,
0.06853367388248444,
0.007354621775448322,
0.08972878009080887,
-0.2669793367385864,
0.019881360232830048,
0.06918594241142273,
-0.0025153355672955513,
0.07059336453676224,
0.06344282627105713,
-0.07033728063106537,
0.10271385312080383,
-0.051166124641895294,
0.1467856466770172,
0.08377711474895477,
-0.09116126596927643,
-0.18892322480678558,
-0.08764564990997314,
0.0990586131811142,
0.17651304602622986,
0.04750865325331688,
-0.024397386237978935,
0.09895956516265869,
-0.0878119245171547,
0.015860557556152344,
0.052259236574172974,
-0.07261253148317337,
-0.05407591536641121,
0.061004482209682465,
0.07816638052463531,
0.06616047024726868,
-0.12551534175872803,
-0.02998468652367592,
0.005221198312938213,
0.011705057695508003,
0.07518111169338226,
0.01836656779050827,
0.15222862362861633,
0.03479425609111786,
-0.12653809785842896,
-0.04834689199924469,
0.0983143299818039,
0.03359128534793854,
-0.043975554406642914,
-0.247073233127594,
-0.031072303652763367,
-0.026882093399763107,
-0.030029185116291046,
-0.038772210478782654,
0.04153512790799141,
-0.006745535880327225,
0.08434242010116577,
-0.0040448750369250774,
-0.07344388216733932,
-0.03874153643846512,
0.06087949126958847,
0.0669754296541214,
0.029331250116229057,
-0.013996441848576069,
0.010876164771616459,
0.11490162461996078,
0.10806918889284134,
-0.12199585139751434,
-0.05589085817337036,
-0.06492951512336731,
-0.08786392956972122,
-0.04284887760877609,
0.033410828560590744,
0.03509693965315819,
0.05435176193714142,
0.2536843419075012,
0.009815474040806293,
0.06126174330711365,
0.03745805472135544,
0.007310505956411362,
0.059651583433151245,
0.10812553018331528,
-0.05987109988927841,
-0.10409316420555115,
-0.02881651371717453,
0.08857584744691849,
0.006609630770981312,
-0.03354408219456673,
-0.05052083358168602,
0.05901389569044113,
0.021856583654880524,
0.11749778687953949,
0.08884359151124954,
0.00984770804643631,
-0.07126569002866745,
-0.06146538630127907,
0.19450126588344574,
-0.16384615004062653,
0.04264351725578308,
0.03702449053525925,
-0.039683789014816284,
-0.0003956064465455711,
0.011445282027125359,
0.01843930408358574,
-0.023893611505627632,
0.09238249063491821,
-0.05498874559998512,
-0.04001082479953766,
-0.1106586754322052,
-0.0339570976793766,
0.034455835819244385,
0.010122774168848991,
-0.03529255837202072,
-0.03252722695469856,
-0.08346389979124069,
-0.07506290078163147,
0.09339368343353271,
-0.07379438728094101,
-0.04854428768157959,
-0.018830472603440285,
-0.0752616599202156,
0.02326788194477558,
0.02032634988427162,
0.07736726850271225,
-0.023358777165412903,
0.04288764297962189,
-0.054010841995477676,
0.05824148654937744,
0.11001134663820267,
0.035365406423807144,
-0.05824809893965721,
0.06025301292538643,
-0.2382364422082901,
0.09637492895126343,
-0.07412451505661011,
0.05830197036266327,
-0.15449334681034088,
-0.02627694234251976,
0.04870045557618141,
0.0076532382518053055,
-0.009597796015441418,
0.13436771929264069,
-0.21578943729400635,
-0.026375943794846535,
0.16865074634552002,
-0.10160042345523834,
-0.06946627050638199,
0.05867103114724159,
-0.049256108701229095,
0.10817171633243561,
0.03891118988394737,
-0.025492025539278984,
0.06244310364127159,
-0.12527504563331604,
0.007147894706577063,
-0.04992884770035744,
-0.016554534435272217,
0.1592475026845932,
0.07294736802577972,
-0.07235062122344971,
0.07110220938920975,
0.025814544409513474,
-0.027441376820206642,
-0.04532165080308914,
-0.016039686277508736,
-0.10585595667362213,
0.014911207370460033,
-0.061168964952230453,
0.01876060478389263,
-0.020111115649342537,
-0.08977947384119034,
-0.028080428019165993,
-0.1748371720314026,
-0.026230180636048317,
0.085477814078331,
-0.007464459165930748,
-0.018854627385735512,
-0.11770102381706238,
0.008567224256694317,
0.044854406267404556,
0.006109896115958691,
-0.13499478995800018,
-0.04764661565423012,
0.027907660230994225,
-0.16220368444919586,
0.033779170364141464,
-0.05184612050652504,
0.05056280270218849,
0.026674345135688782,
-0.029802238568663597,
-0.025906935334205627,
0.022987615317106247,
0.006545235402882099,
-0.011514187790453434,
-0.24465326964855194,
-0.026841215789318085,
-0.026506783440709114,
0.166712686419487,
-0.20777921378612518,
0.03577128052711487,
0.08057375997304916,
0.15318496525287628,
0.011457439512014389,
-0.04087435454130173,
0.005527274217456579,
-0.06868630647659302,
-0.025992877781391144,
-0.05823420733213425,
-0.002480053110048175,
-0.03337050974369049,
-0.04843711107969284,
0.04469521716237068,
-0.1662919819355011,
-0.03491327911615372,
0.09593124687671661,
0.06427760422229767,
-0.13986408710479736,
-0.023568401113152504,
-0.03526119887828827,
-0.049809779971838,
-0.047768235206604004,
-0.06002878025174141,
0.11181395500898361,
0.058611296117305756,
0.04419868439435959,
-0.059296321123838425,
-0.07637067884206772,
-0.0028071242850273848,
-0.014342374168336391,
-0.01986078731715679,
0.097631074488163,
0.06816094368696213,
-0.1381729394197464,
0.09227006882429123,
0.09810956567525864,
0.07738673686981201,
0.09273158758878708,
-0.02444581687450409,
-0.08119411021471024,
-0.0471174530684948,
0.03257923200726509,
0.018235107883810997,
0.1276484578847885,
-0.027872784063220024,
0.04268912971019745,
0.0421174094080925,
-0.018595336005091667,
0.013991083949804306,
-0.08597505837678909,
0.033884208649396896,
0.02703946642577648,
-0.0159194003790617,
0.04745442420244217,
-0.037611253559589386,
0.024539871141314507,
0.08754327148199081,
0.04615016281604767,
0.033831849694252014,
0.015717241913080215,
-0.05243339762091637,
-0.10873834043741226,
0.1642032116651535,
-0.12759798765182495,
-0.22238075733184814,
-0.13922695815563202,
0.003997850697487593,
0.036267586052417755,
-0.01646288111805916,
0.002834152430295944,
-0.060960907489061356,
-0.12132686376571655,
-0.08726011961698532,
0.015815909951925278,
0.050406474620103836,
-0.0912260189652443,
-0.060087788850069046,
0.056193675845861435,
0.037736181169748306,
-0.14546552300453186,
0.01776101253926754,
0.04850281774997711,
-0.09700650721788406,
-0.004754792433232069,
0.07885372638702393,
0.06784981489181519,
0.17673011124134064,
0.018112216144800186,
-0.021776698529720306,
0.031116241589188576,
0.20988549292087555,
-0.13491620123386383,
0.11005933582782745,
0.13349974155426025,
-0.09236859530210495,
0.08153878152370453,
0.20252206921577454,
0.04006611555814743,
-0.09986240416765213,
0.032548144459724426,
0.02142537757754326,
-0.027797512710094452,
-0.2441972941160202,
-0.07161470502614975,
-0.004515932407230139,
-0.06051458790898323,
0.07499068230390549,
0.09190185368061066,
0.08272628486156464,
0.011750337667763233,
-0.09449771046638489,
-0.08492138236761093,
0.06362129002809525,
0.10420511662960052,
0.02181125245988369,
-0.009744768962264061,
0.09036174416542053,
-0.03286943957209587,
0.01948373205959797,
0.08554471284151077,
0.0038120283279567957,
0.18320275843143463,
0.051725953817367554,
0.19073979556560516,
0.07944851368665695,
0.06951095163822174,
0.012023290619254112,
0.011227634735405445,
0.018135491758584976,
0.03228217363357544,
-0.003646562807261944,
-0.08350840210914612,
-0.02080707624554634,
0.1153142973780632,
0.0672341138124466,
0.012952476739883423,
0.01729460060596466,
-0.04021955281496048,
0.08128432929515839,
0.18377035856246948,
-0.0093126455321908,
-0.177269846200943,
-0.06024068966507912,
0.07718996703624725,
-0.09723462164402008,
-0.09738315641880035,
-0.01454379502683878,
0.030975129455327988,
-0.1702532023191452,
0.025819219648838043,
-0.023134231567382812,
0.11114585399627686,
-0.13745717704296112,
-0.020040949806571007,
0.07143081724643707,
0.07336213439702988,
0.004178736824542284,
0.055973317474126816,
-0.16574905812740326,
0.1074945405125618,
0.007851972244679928,
0.06788748502731323,
-0.0949488952755928,
0.10003086179494858,
-0.002759356750175357,
-0.016956903040409088,
0.13766175508499146,
0.003847390878945589,
-0.0742180123925209,
-0.07706846296787262,
-0.08544620126485825,
-0.010016623884439468,
0.12665624916553497,
-0.13990990817546844,
0.08602021634578705,
-0.03789570555090904,
-0.04160536453127861,
-0.0009961887262761593,
-0.09994571655988693,
-0.11771732568740845,
-0.18694964051246643,
0.060274846851825714,
-0.13818500936031342,
0.030693015083670616,
-0.1080726683139801,
-0.033236145973205566,
-0.03044886700809002,
0.18898600339889526,
-0.23496590554714203,
-0.07289838045835495,
-0.14654842019081116,
-0.10314314812421799,
0.14515270292758942,
-0.05135014280676842,
0.0824703797698021,
-0.007518251892179251,
0.16955603659152985,
0.01909777894616127,
-0.024870775640010834,
0.09702518582344055,
-0.09090493619441986,
-0.19369281828403473,
-0.07736486196517944,
0.1553725302219391,
0.13563397526741028,
0.03274888917803764,
-0.0031351360958069563,
0.03731042891740799,
-0.016484085470438004,
-0.119691863656044,
0.016338739544153214,
0.17828133702278137,
0.06005066633224487,
0.02449444867670536,
-0.025351086631417274,
-0.12034450471401215,
-0.07065033912658691,
-0.028268499299883842,
0.030481377616524696,
0.1794593334197998,
-0.06955225765705109,
0.18364831805229187,
0.147920161485672,
-0.05845186114311218,
-0.20284810662269592,
0.01105605997145176,
0.03317207098007202,
-0.00011460785754024982,
0.025185899809002876,
-0.19945523142814636,
0.08448769152164459,
0.004838644526898861,
-0.0498092919588089,
0.1281348466873169,
-0.17351724207401276,
-0.14425379037857056,
0.07726620137691498,
0.03829115256667137,
-0.1926836371421814,
-0.12892304360866547,
-0.09138946235179901,
-0.04540696740150452,
-0.18867050111293793,
0.09461917728185654,
0.031194355338811874,
0.009373899549245834,
0.030387504026293755,
0.030604345723986626,
0.01938873715698719,
-0.04181704297661781,
0.1860174536705017,
-0.023930367082357407,
0.028327496722340584,
-0.08596936613321304,
-0.07190530747175217,
0.0391114242374897,
-0.05227291211485863,
0.07252339273691177,
-0.023452037945389748,
0.00719826715067029,
-0.09769386798143387,
-0.04156304895877838,
-0.03843177855014801,
0.01581472158432007,
-0.09648153930902481,
-0.08523351699113846,
-0.04445706307888031,
0.09780744463205338,
0.09553340077400208,
-0.03473082184791565,
-0.024805041030049324,
-0.07508285343647003,
0.04805302992463112,
0.19605006277561188,
0.17889533936977386,
0.03904116898775101,
-0.07846304774284363,
-0.0033101453445851803,
-0.010484009049832821,
0.04490501061081886,
-0.20383046567440033,
0.06269704550504684,
0.05393069609999657,
0.019165942445397377,
0.11697915196418762,
-0.01937638409435749,
-0.15321338176727295,
-0.07137971371412277,
0.062210626900196075,
-0.05747547000646591,
-0.19925202429294586,
0.008424095809459686,
0.062047190964221954,
-0.16446428000926971,
-0.045800499618053436,
0.046785544604063034,
-0.004990153945982456,
-0.03839265555143356,
0.022938871756196022,
0.09231305122375488,
0.0029900665394961834,
0.07426668703556061,
0.052022483199834824,
0.0835016593337059,
-0.1060708537697792,
0.07922257483005524,
0.08730976283550262,
-0.08381073921918869,
0.022620677947998047,
0.10530175268650055,
-0.061487648636102676,
-0.03560204058885574,
0.017662353813648224,
0.08361397683620453,
0.018624287098646164,
-0.03893670439720154,
0.014383325353264809,
-0.1065717563033104,
0.059272702783346176,
0.08645539730787277,
0.03302672877907753,
0.01618802361190319,
0.034192394465208054,
0.04655340686440468,
-0.06840039044618607,
0.122025266289711,
0.032824426889419556,
0.017204686999320984,
-0.035474274307489395,
-0.04102595895528793,
0.01851540431380272,
-0.03368416428565979,
-0.005532157141715288,
-0.03097093477845192,
-0.07835554331541061,
-0.015077406540513039,
-0.16520504653453827,
-0.009829589165747166,
-0.05936548113822937,
0.012285472825169563,
0.031714752316474915,
-0.034721489995718,
0.008415459655225277,
0.009580436162650585,
-0.07713334262371063,
-0.06541574746370316,
-0.01965213567018509,
0.0961783304810524,
-0.1606777459383011,
0.022340767085552216,
0.08350874483585358,
-0.12098895758390427,
0.09293801337480545,
0.01664864458143711,
-0.00869405921548605,
0.02654755860567093,
-0.1516905426979065,
0.03389517217874527,
-0.03324367105960846,
0.009356614202260971,
0.04251125827431679,
-0.2180858999490738,
-0.0012979574967175722,
-0.034122150391340256,
-0.06511902064085007,
-0.008563618175685406,
-0.035606082528829575,
-0.1133907288312912,
0.10431582480669022,
0.007158213295042515,
-0.08918852359056473,
-0.031932637095451355,
0.02896781638264656,
0.08660420775413513,
-0.02103978954255581,
0.1533614844083786,
-0.008595003746449947,
0.07452014833688736,
-0.16158120334148407,
-0.019116591662168503,
-0.0044966633431613445,
0.021838920190930367,
-0.020337330177426338,
-0.011089952662587166,
0.043057333678007126,
-0.02310733124613762,
0.1769370436668396,
-0.034001484513282776,
0.02080564945936203,
0.06879838556051254,
0.02382824197411537,
-0.03270673379302025,
0.10420172661542892,
0.04176081717014313,
0.020029285922646523,
0.016749408096075058,
0.0014026050921529531,
-0.04661702737212181,
-0.03435906395316124,
-0.1965997964143753,
0.07266207784414291,
0.15759599208831787,
0.09697116911411285,
-0.019108884036540985,
0.07821404188871384,
-0.0993313267827034,
-0.10917975008487701,
0.12915705144405365,
-0.04755320027470589,
-0.004375945311039686,
-0.07154709100723267,
0.13273866474628448,
0.14712604880332947,
-0.18722544610500336,
0.07334931939840317,
-0.07133730500936508,
-0.04749078303575516,
-0.10922681540250778,
-0.194550022482872,
-0.05630992352962494,
-0.049111537635326385,
-0.015855323523283005,
-0.04727233946323395,
0.07431400567293167,
0.05443255603313446,
0.007043207995593548,
-0.0018872307846322656,
0.06250270456075668,
-0.02979675866663456,
-0.004455813206732273,
0.033084239810705185,
0.06524696946144104,
0.012280851602554321,
-0.028982065618038177,
0.017169395461678505,
-0.009704679250717163,
0.04565926641225815,
0.06593092530965805,
0.0490880124270916,
-0.02946917712688446,
0.01301988959312439,
-0.040264759212732315,
-0.10370729863643646,
0.044506072998046875,
-0.02268853597342968,
-0.081757090985775,
0.15341326594352722,
0.023376943543553352,
0.008703592233359814,
-0.018961627036333084,
0.23797030746936798,
-0.07337556779384613,
-0.09915944188833237,
-0.14910556375980377,
0.10603363811969757,
-0.037726908922195435,
0.05897798761725426,
0.04798928648233414,
-0.10144850611686707,
0.018896711990237236,
0.1251462697982788,
0.16306589543819427,
-0.03724272549152374,
0.020064668729901314,
0.030806828290224075,
0.005520908627659082,
-0.035788439214229584,
0.04845234379172325,
0.06755134463310242,
0.16263099014759064,
-0.046816933900117874,
0.09447267651557922,
0.0011601726291701198,
-0.09597980976104736,
-0.03777771443128586,
0.10832508653402328,
-0.014584118500351906,
0.018404638394713402,
-0.059979453682899475,
0.11911186575889587,
-0.06456011533737183,
-0.2371375411748886,
0.062140509486198425,
-0.06866546720266342,
-0.13664314150810242,
-0.023452885448932648,
0.08483598381280899,
-0.011404541321098804,
0.028394777327775955,
0.07356005162000656,
-0.07185159623622894,
0.20126941800117493,
0.03666449710726738,
-0.05399559810757637,
-0.054549336433410645,
0.0827551931142807,
-0.09896446764469147,
0.27000707387924194,
0.015913790091872215,
0.048061735928058624,
0.1041264757514,
-0.008932216092944145,
-0.13759581744670868,
0.019727399572730064,
0.0954047441482544,
-0.10358903557062149,
0.041838936507701874,
0.19829733669757843,
-0.0014832824235782027,
0.1230277270078659,
0.07854447513818741,
-0.07668869197368622,
0.0473078191280365,
-0.08185897022485733,
-0.06852826476097107,
-0.0918748751282692,
0.10061057657003403,
-0.07712632417678833,
0.14169210195541382,
0.13906599581241608,
-0.05018797889351845,
0.011615060269832611,
-0.031394075602293015,
0.04402702674269676,
0.0006254917825572193,
0.10420145094394684,
0.002576707163825631,
-0.18477243185043335,
0.02472778968513012,
0.006634650751948357,
0.10846512019634247,
-0.15925930440425873,
-0.09642539173364639,
0.03936212509870529,
0.004935122560709715,
-0.06595125794410706,
0.1294470727443695,
0.055943287909030914,
0.043614063411951065,
-0.039108045399188995,
-0.036952149122953415,
-0.006302761845290661,
0.13504701852798462,
-0.1053730770945549,
0.002390247769653797
] |
null | null | diffusers |
# DreamBooth - NK2306/FineTunedModelSD
This is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on ku Tenh using [DreamBooth](https://dreambooth.github.io/).
You can find some example images in the following.
DreamBooth for the text encoder was enabled: False.
| {"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "dreambooth"], "base_model": "CompVis/stable-diffusion-v1-4", "instance_prompt": "ku Tenh", "inference": true} | text-to-image | NK2306/FineTunedModelSD | [
"diffusers",
"tensorboard",
"safetensors",
"stable-diffusion",
"stable-diffusion-diffusers",
"text-to-image",
"dreambooth",
"base_model:CompVis/stable-diffusion-v1-4",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-06T04:43:23+00:00 | [] | [] | TAGS
#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
|
# DreamBooth - NK2306/FineTunedModelSD
This is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on ku Tenh using DreamBooth.
You can find some example images in the following.
DreamBooth for the text encoder was enabled: False.
| [
"# DreamBooth - NK2306/FineTunedModelSD\n\nThis is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on ku Tenh using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
"TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n",
"# DreamBooth - NK2306/FineTunedModelSD\n\nThis is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on ku Tenh using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
97,
79
] | [
"passage: TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #dreambooth #base_model-CompVis/stable-diffusion-v1-4 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n# DreamBooth - NK2306/FineTunedModelSD\n\nThis is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on ku Tenh using DreamBooth.\nYou can find some example images in the following. \n\n\n\nDreamBooth for the text encoder was enabled: False."
] | [
-0.042480554431676865,
0.06309319287538528,
-0.0009013091330416501,
0.0314345583319664,
0.11230699717998505,
-0.0009034625254571438,
0.20968398451805115,
0.028753286227583885,
0.058545880019664764,
0.0639677569270134,
0.10546842217445374,
-0.02398792840540409,
0.023897044360637665,
0.11465507000684738,
0.026531951501965523,
-0.16817082464694977,
0.026981176808476448,
0.009436839260160923,
-0.009138382039964199,
0.042563654482364655,
0.02708243578672409,
-0.10223852843046188,
0.09533882141113281,
-0.03601051867008209,
-0.17115333676338196,
0.01369374431669712,
-0.013472416438162327,
-0.04549907147884369,
0.06241225823760033,
0.019309690222144127,
0.12685707211494446,
0.08440246433019638,
0.01707381196320057,
-0.1518261432647705,
0.031494803726673126,
0.04360023885965347,
-0.002341795712709427,
0.024157097563147545,
-0.0716988816857338,
0.023683389648795128,
0.08769384026527405,
-0.08091771602630615,
0.05592918023467064,
0.005446793511509895,
-0.060655608773231506,
0.10757263004779816,
-0.0063944351859390736,
0.08729603886604309,
0.0731627345085144,
0.03667774796485901,
-0.020590968430042267,
0.0486329086124897,
0.006497248075902462,
0.09203694760799408,
0.2046400010585785,
-0.1882735639810562,
-0.05943836644291878,
0.2519115209579468,
0.03864069655537605,
0.022273730486631393,
-0.02951575629413128,
0.05444954335689545,
0.06538848578929901,
0.010154756717383862,
0.05908333882689476,
-0.07894663512706757,
-0.049754705280065536,
-0.13244274258613586,
-0.10340949892997742,
0.03164564073085785,
0.05636122450232506,
-0.012920175679028034,
-0.04316193610429764,
-0.1698143482208252,
-0.08311615884304047,
0.09073293209075928,
-0.008318407461047173,
-0.014185305684804916,
-0.03615204989910126,
-0.04643433913588524,
-0.05585947632789612,
-0.022655485197901726,
-0.04898283630609512,
-0.04192347079515457,
0.033975083380937576,
0.10773495584726334,
0.026050705462694168,
0.0013298451667651534,
-0.04452410712838173,
0.1483333259820938,
-0.061058785766363144,
-0.14951252937316895,
0.052060116082429886,
-0.08812501281499863,
0.02197825163602829,
0.06732001900672913,
0.015615622512996197,
-0.21431270241737366,
0.08174961060285568,
-0.01726163737475872,
0.08807004988193512,
0.022680670022964478,
0.028652207925915718,
0.03507780656218529,
-0.0002020709216594696,
0.019682519137859344,
-0.02856176160275936,
-0.088262639939785,
0.015180733054876328,
0.06953290104866028,
0.04190864413976669,
-0.047444842755794525,
-0.08151152729988098,
0.03340166062116623,
-0.019218558445572853,
0.050263796001672745,
-0.03475981205701828,
0.015776067972183228,
-0.0724838450551033,
-0.015457632951438427,
0.08630522340536118,
-0.03295085206627846,
0.0030538863502442837,
-0.059926845133304596,
0.012581498362123966,
0.035846833139657974,
0.13007456064224243,
0.016537532210350037,
-0.0526193343102932,
0.091216541826725,
-0.09182505309581757,
0.024660624563694,
-0.008724560961127281,
-0.10379244387149811,
0.010377127677202225,
-0.3001443147659302,
0.028217654675245285,
-0.1316549926996231,
-0.09271501004695892,
-0.013792885467410088,
0.030176782980561256,
0.032874930649995804,
0.026522668078541756,
-0.08371400833129883,
-0.09446075558662415,
-0.03578836843371391,
0.06023775786161423,
-0.04898672550916672,
0.014206133782863617,
0.00978136993944645,
-0.04001430422067642,
0.11084531992673874,
-0.044460054486989975,
-0.03633842617273331,
-0.09984081238508224,
0.03809549659490585,
-0.06495899707078934,
0.08627232164144516,
-0.0751805379986763,
0.0743676945567131,
-0.039875440299510956,
-0.026212332770228386,
0.036937762051820755,
-0.004213767126202583,
0.06959269195795059,
0.1423741728067398,
-0.21337313950061798,
-0.05372357740998268,
0.18952536582946777,
-0.15082009136676788,
-0.11320014297962189,
0.03929521143436432,
-0.012777506373822689,
0.09620513021945953,
0.0947379544377327,
0.12819364666938782,
0.10411782562732697,
-0.2838515043258667,
-0.05623433738946915,
-0.0761803612112999,
-0.029924312606453896,
0.015615325421094894,
-0.018467796966433525,
0.08850555121898651,
-0.019552187994122505,
0.042070258408784866,
-0.01910744048655033,
0.09449200332164764,
-0.043328266590833664,
-0.029070263728499413,
-0.044532034546136856,
-0.07627303153276443,
0.03704750910401344,
-0.015038097277283669,
0.023373844102025032,
0.0019746229518204927,
-0.012784013524651527,
0.08290029317140579,
0.03530518710613251,
-0.08212820440530777,
0.00193941465113312,
-0.05255106836557388,
0.0342114120721817,
-0.08674127608537674,
-0.009078009985387325,
-0.08612701296806335,
-0.037864550948143005,
0.05470145493745804,
0.07416913658380508,
0.008856590837240219,
0.031324975192546844,
0.047140300273895264,
0.12332800030708313,
-0.019936177879571915,
-0.049500491470098495,
0.020141486078500748,
0.047566354274749756,
-0.04559006541967392,
-0.14400264620780945,
0.03685682639479637,
-0.0725422203540802,
-0.053448792546987534,
-0.14859594404697418,
0.06008421629667282,
0.03307857736945152,
0.22735022008419037,
0.08429156243801117,
-0.040622860193252563,
0.06862335652112961,
0.0604398138821125,
-0.004722535144537687,
-0.10435105115175247,
0.004776785150170326,
-0.00868369173258543,
-0.12962020933628082,
0.10408831387758255,
-0.13450482487678528,
0.043357543647289276,
0.08668581396341324,
0.1727062612771988,
-0.003559362841770053,
0.07859603315591812,
-0.04054062068462372,
-0.02314247563481331,
-0.06651551276445389,
0.04992431402206421,
0.12060379236936569,
0.00642079534009099,
0.11790218204259872,
-0.028798958286643028,
0.03816871717572212,
0.10375316441059113,
-0.02747727558016777,
-0.04894046112895012,
0.0801817774772644,
-0.040913455188274384,
-0.01712372899055481,
0.004961684811860323,
0.07406681030988693,
0.02690224163234234,
0.2348250299692154,
-0.023309262469410896,
-0.01885095238685608,
-0.08207011222839355,
-0.018366508185863495,
0.018486015498638153,
0.2014952152967453,
-0.033141497522592545,
-0.02512996457517147,
-0.027642732486128807,
-0.03273675590753555,
0.013177887536585331,
-0.15912598371505737,
-0.019030068069696426,
0.010487977415323257,
-0.019173096865415573,
0.14039568603038788,
0.07176369428634644,
-0.1289597898721695,
0.018191589042544365,
-0.11435388028621674,
-0.01311403512954712,
0.010309897363185883,
-0.02404727227985859,
-0.05941137298941612,
0.1392395794391632,
-0.0790516585111618,
-0.2370157092809677,
-0.09925305098295212,
0.02858676202595234,
-0.03137568384408951,
-0.020887341350317,
0.04434092715382576,
-0.06377881020307541,
-0.03701122850179672,
-0.11219505965709686,
0.031194789335131645,
0.07613922655582428,
0.042736396193504333,
0.06450827419757843,
-0.017580030485987663,
-0.0013594693737104535,
-0.047440215945243835,
-0.015451623126864433,
-0.05194926634430885,
0.015437395311892033,
0.07143574208021164,
0.013305816799402237,
0.10775455832481384,
0.14231513440608978,
0.000959125638473779,
-0.018337739631533623,
0.0014808675041422248,
0.20237195491790771,
-0.0005088815814815462,
0.0876372903585434,
0.11801297217607498,
0.020351536571979523,
0.06824410706758499,
0.14294147491455078,
0.03539840877056122,
-0.03092499077320099,
0.0968674048781395,
-0.01201652828603983,
-0.11922893673181534,
-0.07487645745277405,
-0.06937241554260254,
0.001666757627390325,
0.030875571072101593,
0.055469874292612076,
0.018154820427298546,
0.07523659616708755,
0.11130556464195251,
0.05452193692326546,
0.10166318714618683,
0.033698417246341705,
0.08121807128190994,
-0.015944652259349823,
-0.06025150790810585,
0.018139589577913284,
-0.0655452162027359,
-0.06816324591636658,
0.03549953177571297,
-0.05918502435088158,
0.15179777145385742,
-0.07979413121938705,
-0.06884018331766129,
0.0906096026301384,
0.034193672239780426,
0.11689794063568115,
0.07135006040334702,
-0.01450374349951744,
-0.033958375453948975,
-0.015624510124325752,
-0.09123098850250244,
0.02746519073843956,
0.0569412037730217,
-0.03289163485169411,
-0.0073643275536596775,
-0.014518349431455135,
0.16009187698364258,
0.020338507369160652,
0.02034546434879303,
0.16952262818813324,
-0.2096223682165146,
-0.04315977543592453,
-0.002509429818019271,
0.013572068884968758,
-0.0326136015355587,
0.011148663237690926,
0.3134068250656128,
-0.06755252182483673,
0.05701703950762749,
-0.09628891199827194,
0.03491193428635597,
0.006785496138036251,
-0.005784275010228157,
-0.09982136636972427,
0.0839754194021225,
-0.04217896983027458,
-0.014870060607790947,
-0.3104863166809082,
0.019513849169015884,
-0.03183584287762642,
0.11103501915931702,
-0.04262881726026535,
-0.020537642762064934,
0.02436639554798603,
0.04484568163752556,
0.11820561438798904,
0.037021514028310776,
0.04574155807495117,
-0.05304461717605591,
-0.14979787170886993,
0.02750110626220703,
0.040106967091560364,
-0.01994941011071205,
0.05091502517461777,
0.1055155023932457,
-0.026292581111192703,
-0.004241389222443104,
-0.017115889117121696,
-0.24169737100601196,
-0.06401363015174866,
0.001211759983561933,
0.13134726881980896,
0.008008414879441261,
-0.09082339704036713,
-0.09579093009233475,
0.08684419840574265,
0.02448030561208725,
-0.17473123967647552,
-0.09924302995204926,
-0.06078184396028519,
-0.060951631516218185,
0.08066800236701965,
-0.055744484066963196,
0.05661863833665848,
-0.013907460495829582,
0.07240679860115051,
-0.12675794959068298,
-0.090377576649189,
0.02622770518064499,
-0.10420794039964676,
-0.1772773414850235,
-0.09908932447433472,
0.062316883355379105,
0.05423467606306076,
-0.02698446437716484,
-0.02845609001815319,
-0.00041624688310548663,
0.03914147615432739,
-0.07055624574422836,
0.019946226850152016,
0.23476305603981018,
-0.10425583273172379,
0.07460319995880127,
0.024709800258278847,
-0.046586185693740845,
-0.08347698301076889,
0.026295267045497894,
0.05235753208398819,
0.2002619057893753,
-0.032470330595970154,
0.12309373915195465,
0.15155334770679474,
-0.0902375653386116,
-0.251761794090271,
-0.09333224594593048,
0.016470326110720634,
0.05290919914841652,
-0.017978081479668617,
-0.13618704676628113,
0.14315561950206757,
-0.052792247384786606,
0.006888394709676504,
0.10219387710094452,
-0.3487965762615204,
-0.1446208953857422,
-0.0021350006572902203,
0.2013430893421173,
0.2403857409954071,
-0.06531687080860138,
-0.028161877766251564,
0.005818809382617474,
-0.11273065954446793,
0.21554486453533173,
-0.10431433469057083,
0.0768556147813797,
-0.05041243135929108,
-0.03255956992506981,
0.0011014582123607397,
-0.05551522970199585,
0.062058962881565094,
0.010063549503684044,
0.028719931840896606,
-0.06640294194221497,
-0.03255078196525574,
0.19187702238559723,
-0.09309045225381851,
0.041084323078393936,
-0.11456983536481857,
0.06230840086936951,
-0.03558836132287979,
-0.014221752993762493,
-0.022115787491202354,
0.01259087584912777,
-0.04939309135079384,
-0.10976757109165192,
-0.030204972252249718,
0.041351523250341415,
0.07916029542684555,
0.011316648684442043,
-0.01860782504081726,
0.0546187199652195,
-0.038703802973032,
0.11368570476770401,
-0.02563748136162758,
-0.026361431926488876,
-0.03459487110376358,
-0.012702257372438908,
-0.038085393607616425,
0.11263930052518845,
-0.11980073899030685,
-0.03222596272826195,
0.14662599563598633,
0.07052996754646301,
0.13258007168769836,
0.0234905406832695,
-0.12054722011089325,
0.04013923928141594,
0.08281060308218002,
-0.1588851362466812,
-0.05480483919382095,
-0.04370392858982086,
-0.07002459466457367,
0.05762539803981781,
0.004769950173795223,
0.16761572659015656,
-0.14584459364414215,
0.01891387067735195,
-0.045589786022901535,
0.04593885317444801,
-0.03129170835018158,
0.11422885954380035,
0.021050743758678436,
0.05234946683049202,
-0.05528833717107773,
0.10105481743812561,
-0.015867868438363075,
-0.1134677454829216,
0.04030594229698181,
0.020100219175219536,
-0.1464492678642273,
0.0007161052781157196,
-0.0411514975130558,
0.18575388193130493,
-0.030500827357172966,
-0.06876978278160095,
-0.09830465167760849,
-0.09477603435516357,
-0.004483731929212809,
0.14496353268623352,
0.0227228831499815,
0.06361696124076843,
-0.05569448322057724,
-0.03951563686132431,
-0.08890477567911148,
0.10650848597288132,
0.0772460326552391,
0.08520622551441193,
-0.21825982630252838,
-0.011647873558104038,
0.04418855905532837,
0.02972385846078396,
-0.07037863880395889,
-0.013467998243868351,
-0.07808010280132294,
-0.0302064660936594,
-0.038858573883771896,
0.11085738986730576,
-0.09761819988489151,
-0.04913339018821716,
-0.03106814995408058,
-0.004157682415097952,
0.0025183686520904303,
0.07425953447818756,
0.012023239396512508,
0.009548219852149487,
-0.017507605254650116,
-0.005932928994297981,
-0.022339092567563057,
-0.07177592813968658,
0.0009049710351973772,
-0.10002335160970688,
0.04085002839565277,
-0.06357724219560623,
-0.09077528119087219,
0.03589136525988579,
-0.174773171544075,
0.041658636182546616,
0.12313807010650635,
-0.028610236942768097,
-0.011106173507869244,
-0.007069575600326061,
-0.045164864510297775,
-0.027810772880911827,
0.05117205157876015,
-0.014996211044490337,
0.017549952492117882,
-0.07055781036615372,
-0.05416562780737877,
0.01433009747415781,
0.048739586025476456,
-0.06017480418086052,
0.04969467967748642,
0.08673228323459625,
0.12414827197790146,
0.13072480261325836,
-0.1319587528705597,
0.10016097128391266,
-0.07968806475400925,
-0.03385727107524872,
0.030482813715934753,
-0.0615353062748909,
0.0329403430223465,
-0.016124924644827843,
-0.017255978658795357,
-0.02365897037088871,
0.12445229291915894,
0.025981208309531212,
-0.14503899216651917,
-0.030174707993865013,
-0.03523673862218857,
-0.06551989912986755,
0.04148025065660477,
0.20907482504844666,
-0.008606983348727226,
-0.002040264895185828,
-0.15083976089954376,
0.037054724991321564,
0.08641529828310013,
0.1329944133758545,
0.07769691944122314,
0.03244972601532936,
0.06497122347354889,
0.09949927031993866,
0.030853774398565292,
0.09727279841899872,
-0.01012453529983759,
0.07139986008405685,
-0.1590714454650879,
0.14526993036270142,
-0.0491703525185585,
-0.01296345703303814,
0.04165882617235184,
-0.04746087267994881,
-0.055254094302654266,
0.10557854175567627,
-0.09725083410739899,
-0.029733790084719658,
-0.12172987312078476,
-0.04689120128750801,
-0.08050397038459778,
0.042751848697662354,
-0.07786395400762558,
-0.024937035515904427,
0.042276956140995026,
0.02480621263384819,
0.022998714819550514,
0.19390979409217834,
0.07300886511802673,
-0.011066911742091179,
0.1647125631570816,
-0.006998737808316946,
-0.09081883728504181,
0.0645933449268341,
0.05419237166643143,
0.015999166294932365,
0.043340206146240234,
0.004015783313661814,
0.04616508632898331,
0.011088450439274311,
0.045683085918426514,
0.05575227737426758,
-0.06557569652795792,
-0.004483684431761503,
-0.025083603337407112,
0.0029703511390835047,
0.07741330564022064,
0.11610575020313263,
-0.034608036279678345,
-0.018724553287029266,
0.12369344383478165,
-0.02432415448129177,
-0.1335146129131317,
-0.12388186156749725,
0.08037512749433517,
-0.05955027416348457,
0.03731254115700722,
-0.019222386181354523,
-0.09284818172454834,
-0.03732352331280708,
0.11264150589704514,
0.1298120617866516,
0.027231888845562935,
0.017506692558526993,
-0.0848155990242958,
-0.008636212907731533,
-0.07010239362716675,
0.09175697714090347,
-0.027033688500523567,
0.20865891873836517,
-0.04996507614850998,
0.01526105310767889,
-0.08160677552223206,
-0.1464477926492691,
-0.059527356177568436,
-0.17597030103206635,
0.01941915974020958,
-0.05437159538269043,
-0.06990916281938553,
0.042567625641822815,
-0.19334398210048676,
-0.13259093463420868,
0.20436687767505646,
-0.1456354707479477,
-0.02845354564487934,
-0.05404131859540939,
0.12766510248184204,
0.033209528774023056,
0.04692555591464043,
-0.045724883675575256,
0.04587724432349205,
0.10081776231527328,
-0.01846708543598652,
-0.06043076515197754,
0.08312136679887772,
-0.0758916586637497,
-0.20498552918434143,
0.1108575090765953,
-0.01865863986313343,
0.09788192808628082,
0.04419108107686043,
0.0058847451582551,
-0.06725161522626877,
0.061130862683057785,
-0.04109099879860878,
-0.00019416138820815831,
-0.0974564254283905,
0.1728621870279312,
0.017038604244589806,
0.029345478862524033,
0.016915561631321907,
-0.11030204594135284,
-0.015642983838915825,
0.09408774971961975,
-0.0013183339033275843,
-0.11497307568788528,
0.01315564475953579,
-0.023214085027575493,
0.11346666514873505,
0.0747833326458931,
-0.0664573386311531,
0.05840916186571121,
0.0018308935686945915,
-0.007456247694790363,
-0.023543410003185272,
0.031162617728114128,
0.07459431141614914,
-0.11924951523542404,
0.002730760257691145,
-0.03433539345860481,
0.025998855009675026,
-0.24299189448356628,
-0.06548799574375153,
-0.13826745748519897,
-0.019823377951979637,
-0.03664543479681015,
0.09644850343465805,
0.18110480904579163,
0.04759340360760689,
0.008106330409646034,
-0.16235478222370148,
0.005783341359347105,
0.07899178564548492,
-0.04583066701889038,
-0.10733838379383087
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-classification | dylansheils0241/Quantum-Balanced-GPT2-Experimental-Theoretical-Classifier-Arxiv | [
"transformers",
"safetensors",
"gpt2",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T04:52:14+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #gpt2 #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #gpt2 #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
57,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #gpt2 #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.056688208132982254,
0.16275818645954132,
-0.004995849449187517,
0.02452681213617325,
0.09786820411682129,
0.012234015390276909,
0.0665494054555893,
0.11446622014045715,
-0.019654832780361176,
0.11005201935768127,
0.03224760666489601,
0.09871180355548859,
0.11353186517953873,
0.16168983280658722,
0.0032966197468340397,
-0.23542851209640503,
0.0468909777700901,
-0.12772662937641144,
-0.03536919504404068,
0.1168198511004448,
0.1453043669462204,
-0.10465475916862488,
0.07777879387140274,
-0.030737513676285744,
-0.00966490525752306,
-0.031576599925756454,
-0.06401672959327698,
-0.05254754796624184,
0.05026170611381531,
0.07030009478330612,
0.06729904562234879,
0.007110965438187122,
0.09183509647846222,
-0.26528099179267883,
0.02310207113623619,
0.07594640552997589,
-0.001530517591163516,
0.07526049017906189,
0.04812141880393028,
-0.08188801258802414,
0.07724089175462723,
-0.05910906940698624,
0.14970065653324127,
0.08168394863605499,
-0.08909645676612854,
-0.19279919564723969,
-0.08510304987430573,
0.10391399264335632,
0.18374687433242798,
0.05092503875494003,
-0.023054618388414383,
0.10542254149913788,
-0.08633118867874146,
0.009634695015847683,
0.05877980589866638,
-0.06533382833003998,
-0.05344819277524948,
0.07030105590820312,
0.07782267779111862,
0.07990992069244385,
-0.1312698870897293,
-0.015684951096773148,
0.011104214005172253,
0.00718331104144454,
0.08438988029956818,
0.021947048604488373,
0.13984841108322144,
0.04317888617515564,
-0.13002023100852966,
-0.043909087777137756,
0.09818204492330551,
0.04373921453952789,
-0.05486512929201126,
-0.2536156177520752,
-0.024169430136680603,
-0.04035896062850952,
-0.030552461743354797,
-0.03810810670256615,
0.046126991510391235,
-0.011335964314639568,
0.08135324716567993,
-0.01014806516468525,
-0.08107352256774902,
-0.039619747549295425,
0.06577171385288239,
0.06208868697285652,
0.026821019127964973,
-0.011671994812786579,
0.010234266519546509,
0.12182370573282242,
0.11196604371070862,
-0.1267545372247696,
-0.057587433606386185,
-0.06730430573225021,
-0.08423306047916412,
-0.0447598397731781,
0.033019017428159714,
0.043012093752622604,
0.043217264115810394,
0.24644720554351807,
0.009206846356391907,
0.056639526039361954,
0.046275459229946136,
0.01094451081007719,
0.06766996532678604,
0.11699596792459488,
-0.06068361550569534,
-0.09169819205999374,
-0.028210563585162163,
0.09167566895484924,
0.00641299644485116,
-0.04253393039107323,
-0.06145499274134636,
0.064153291285038,
0.012870057485997677,
0.12154073268175125,
0.08118180185556412,
0.003567382926121354,
-0.07309924066066742,
-0.06970501691102982,
0.1891379952430725,
-0.16031186282634735,
0.048485323786735535,
0.031144030392169952,
-0.04012530297040939,
-0.0037560793571174145,
0.010843836702406406,
0.023869646713137627,
-0.02002854458987713,
0.09128747135400772,
-0.05557823181152344,
-0.03632063418626785,
-0.11016727983951569,
-0.032307349145412445,
0.031602974981069565,
0.005229081027209759,
-0.027400631457567215,
-0.034758999943733215,
-0.088483065366745,
-0.05970429629087448,
0.09185574948787689,
-0.0725538581609726,
-0.05133562162518501,
-0.014709736220538616,
-0.07693814486265182,
0.023501750081777573,
0.017334258183836937,
0.08247776329517365,
-0.02171487919986248,
0.03736530989408493,
-0.05768737196922302,
0.0629277229309082,
0.11147940158843994,
0.028959784656763077,
-0.053054340183734894,
0.05642477050423622,
-0.24564996361732483,
0.10391399264335632,
-0.07363908737897873,
0.050764378160238266,
-0.1495596021413803,
-0.026108751073479652,
0.04620397463440895,
0.008229508064687252,
-0.013179546222090721,
0.14200609922409058,
-0.2165803462266922,
-0.03158329799771309,
0.1675405651330948,
-0.09414954483509064,
-0.07587666064500809,
0.058981720358133316,
-0.05404539033770561,
0.10750386118888855,
0.044090528041124344,
-0.02138153277337551,
0.06216038018465042,
-0.14397647976875305,
-0.012432611547410488,
-0.041649285703897476,
-0.02405562996864319,
0.16330546140670776,
0.07457496225833893,
-0.06195339933037758,
0.08059893548488617,
0.02354867197573185,
-0.01621856912970543,
-0.04427330568432808,
-0.022748108953237534,
-0.10862526297569275,
0.010284756310284138,
-0.060920849442481995,
0.025102287530899048,
-0.024612700566649437,
-0.09277556091547012,
-0.029063016176223755,
-0.17864122986793518,
-0.006479884497821331,
0.0886693224310875,
-0.011500544846057892,
-0.021484078839421272,
-0.12116942554712296,
0.009813597425818443,
0.03316587954759598,
0.0017007269198074937,
-0.13188430666923523,
-0.05852383002638817,
0.027688656002283096,
-0.16301771998405457,
0.03372528776526451,
-0.05495554581284523,
0.05014612898230553,
0.03457905724644661,
-0.03310814127326012,
-0.0332946851849556,
0.009236995130777359,
0.006859376560896635,
-0.012469877488911152,
-0.2507966458797455,
-0.023342570289969444,
-0.022434204816818237,
0.1754746288061142,
-0.22036518156528473,
0.04243079200387001,
0.07755454629659653,
0.1488277018070221,
0.00758343143388629,
-0.0385059118270874,
0.0108621371909976,
-0.081815205514431,
-0.030218740925192833,
-0.06505976617336273,
-0.012460521422326565,
-0.03667744621634483,
-0.05610157549381256,
0.0502706803381443,
-0.15897555649280548,
-0.02874920517206192,
0.1054336428642273,
0.06867773085832596,
-0.14192140102386475,
-0.019301993772387505,
-0.04151778295636177,
-0.043261781334877014,
-0.059273406863212585,
-0.052341289818286896,
0.12165731191635132,
0.05511542409658432,
0.04830443114042282,
-0.06941264122724533,
-0.07721688598394394,
0.007358690723776817,
-0.01781853660941124,
-0.01777339167892933,
0.0893767699599266,
0.07364588975906372,
-0.1244131475687027,
0.09259122610092163,
0.10102657228708267,
0.09181178361177444,
0.1081208810210228,
-0.018081961199641228,
-0.08706964552402496,
-0.044923022389411926,
0.0203016996383667,
0.013586522080004215,
0.14873908460140228,
-0.013833221048116684,
0.056480638682842255,
0.037942372262477875,
-0.011916344054043293,
0.011820046231150627,
-0.09332224726676941,
0.02717048116028309,
0.03382128104567528,
-0.01951325498521328,
0.0386216901242733,
-0.041371311992406845,
0.02094746008515358,
0.08781760185956955,
0.04612930491566658,
0.043133337050676346,
0.0117363091558218,
-0.04232318326830864,
-0.10922972857952118,
0.17441187798976898,
-0.1265207678079605,
-0.24811580777168274,
-0.14101167023181915,
0.003548931796103716,
0.04161277040839195,
-0.00981870200484991,
0.0033757472410798073,
-0.07208114862442017,
-0.11677295714616776,
-0.09369951486587524,
0.014811713248491287,
0.04634920507669449,
-0.07345830649137497,
-0.059883084148168564,
0.06119617074728012,
0.03956949710845947,
-0.1438819319009781,
0.022251952439546585,
0.04618961736559868,
-0.08916490525007248,
-0.007660253439098597,
0.08383868634700775,
0.0663260892033577,
0.17821210622787476,
0.006946447771042585,
-0.0211239792406559,
0.030622977763414383,
0.21115799248218536,
-0.14198173582553864,
0.11189346015453339,
0.13871541619300842,
-0.0914137065410614,
0.08057228475809097,
0.19499939680099487,
0.039358362555503845,
-0.10546562820672989,
0.03305496647953987,
0.025585372000932693,
-0.028574291616678238,
-0.25104397535324097,
-0.0688476711511612,
0.0032778505701571703,
-0.04926806315779686,
0.07188434898853302,
0.09186343103647232,
0.0980982705950737,
0.014891031198203564,
-0.0960264801979065,
-0.08064629137516022,
0.047425657510757446,
0.1060817539691925,
0.0058129834942519665,
-0.014884848147630692,
0.08918209373950958,
-0.033973969519138336,
0.01830443926155567,
0.09493532031774521,
0.0027011006604880095,
0.1717296540737152,
0.05488133803009987,
0.17306490242481232,
0.07624487578868866,
0.06581548601388931,
0.01999000459909439,
0.005975092761218548,
0.02183385193347931,
0.017030013725161552,
-0.0042428807355463505,
-0.08410344272851944,
-0.006547621451318264,
0.11983280628919601,
0.06439385563135147,
0.025222649797797203,
0.01179569773375988,
-0.03775259107351303,
0.08442632108926773,
0.1749287247657776,
0.0019723710138350725,
-0.18502026796340942,
-0.07310080528259277,
0.07859374582767487,
-0.09147641062736511,
-0.10472535341978073,
-0.033703628927469254,
0.03427765890955925,
-0.17517809569835663,
0.022020921111106873,
-0.017561664804816246,
0.1075635626912117,
-0.14025665819644928,
-0.018815070390701294,
0.06322184205055237,
0.07119301706552505,
-0.003150238422676921,
0.06206432729959488,
-0.15356871485710144,
0.10289750248193741,
0.01623719558119774,
0.06722601503133774,
-0.09720844030380249,
0.10133114457130432,
-0.004952651914209127,
-0.004525967873632908,
0.13827085494995117,
0.01132762711495161,
-0.05856676027178764,
-0.09017363935709,
-0.10513435304164886,
-0.00838832464069128,
0.12849153578281403,
-0.15005145967006683,
0.08388647437095642,
-0.02834545075893402,
-0.04350460320711136,
0.0028875069692730904,
-0.12221316248178482,
-0.13001364469528198,
-0.18832695484161377,
0.056929487735033035,
-0.13712386786937714,
0.03853238746523857,
-0.10552185028791428,
-0.04352074861526489,
-0.02887987159192562,
0.20707611739635468,
-0.23298580944538116,
-0.06692921370267868,
-0.15351387858390808,
-0.08341002464294434,
0.14223980903625488,
-0.047327857464551926,
0.08947230875492096,
-0.0006640679202973843,
0.18996402621269226,
0.023749830201268196,
-0.022958843037486076,
0.1032608300447464,
-0.0953274741768837,
-0.19397945702075958,
-0.08692745119333267,
0.1546938568353653,
0.1391226202249527,
0.03686843439936638,
-0.004141106735914946,
0.02847231552004814,
-0.016847684979438782,
-0.124631866812706,
0.024518294259905815,
0.18073658645153046,
0.08076948672533035,
0.0265391543507576,
-0.027251770719885826,
-0.11268990486860275,
-0.06502450257539749,
-0.03222431242465973,
0.02961815521121025,
0.18243654072284698,
-0.07519686967134476,
0.1901955008506775,
0.13443516194820404,
-0.05532145872712135,
-0.19614705443382263,
0.020869113504886627,
0.040682703256607056,
0.010022376663982868,
0.03179483115673065,
-0.20703792572021484,
0.09156925976276398,
0.0010500495554879308,
-0.05144895613193512,
0.13911594450473785,
-0.17432743310928345,
-0.151635080575943,
0.06508468836545944,
0.03710223361849785,
-0.1928347796201706,
-0.11609705537557602,
-0.08698900043964386,
-0.04764305055141449,
-0.17105866968631744,
0.10657104849815369,
0.01810937561094761,
0.00958965066820383,
0.03534632921218872,
0.024905970320105553,
0.011848334223031998,
-0.038653843104839325,
0.1937946081161499,
-0.025732165202498436,
0.030279111117124557,
-0.08093483000993729,
-0.06396064907312393,
0.05981128662824631,
-0.05653243884444237,
0.07926622778177261,
-0.019695239141583443,
0.013051961548626423,
-0.11097230017185211,
-0.04620685055851936,
-0.0259378794580698,
0.01774062030017376,
-0.09204115718603134,
-0.09366398304700851,
-0.052516400814056396,
0.09481028467416763,
0.08910299837589264,
-0.03606134280562401,
-0.037287481129169464,
-0.07352357357740402,
0.033029962331056595,
0.1711474508047104,
0.17759248614311218,
0.03955881670117378,
-0.08047898858785629,
-0.0056870789267122746,
-0.012103362940251827,
0.04403502494096756,
-0.22654037177562714,
0.06325183808803558,
0.04530617222189903,
0.01543486025184393,
0.11385229974985123,
-0.01837165653705597,
-0.1614435911178589,
-0.06635874509811401,
0.06131202355027199,
-0.06802738457918167,
-0.18228155374526978,
0.003949065692722797,
0.059882380068302155,
-0.16518694162368774,
-0.03804792836308479,
0.04252162575721741,
-0.005406341515481472,
-0.04330739378929138,
0.016907572746276855,
0.08069808781147003,
0.005998523905873299,
0.07098331302404404,
0.05882091447710991,
0.08421110361814499,
-0.1037181094288826,
0.07749924808740616,
0.08153212815523148,
-0.08422624319791794,
0.03190286085009575,
0.08996295183897018,
-0.06251151114702225,
-0.031116170808672905,
0.03407644107937813,
0.07890257239341736,
0.021274501457810402,
-0.042060818523168564,
0.007018920499831438,
-0.09714305400848389,
0.06359373033046722,
0.097147636115551,
0.03615867346525192,
0.014581901021301746,
0.033455245196819305,
0.04942217841744423,
-0.07525192946195602,
0.11710844188928604,
0.022087544202804565,
0.018151897937059402,
-0.04620935395359993,
-0.053078390657901764,
0.03228947892785072,
-0.022450316697359085,
-0.00941105093806982,
-0.03882573917508125,
-0.07127608358860016,
-0.018803894519805908,
-0.1570902317762375,
-0.012990092858672142,
-0.04015745967626572,
0.008910941891372204,
0.025404607877135277,
-0.036834996193647385,
0.00830900203436613,
0.01183675043284893,
-0.070340096950531,
-0.06075016036629677,
-0.022623594850301743,
0.09379690140485764,
-0.16431091725826263,
0.02377411536872387,
0.08262428641319275,
-0.12145231664180756,
0.09788227826356888,
0.02060251496732235,
0.0030464455485343933,
0.027605198323726654,
-0.15334390103816986,
0.04065665975213051,
-0.02279796451330185,
0.012528573162853718,
0.04368005692958832,
-0.22050756216049194,
0.0022407027427107096,
-0.03955530747771263,
-0.0581013448536396,
-0.009252764284610748,
-0.02985531836748123,
-0.11230431497097015,
0.09915203601121902,
0.008330761454999447,
-0.08242792636156082,
-0.031244277954101562,
0.03579036891460419,
0.08468461036682129,
-0.025901922956109047,
0.15200486779212952,
0.0009069927618838847,
0.07067206501960754,
-0.17469091713428497,
-0.02179952897131443,
-0.011881532147526741,
0.022119969129562378,
-0.014708246104419231,
-0.015323184430599213,
0.0425114743411541,
-0.024193845689296722,
0.1896541714668274,
-0.024643851444125175,
0.03960045427083969,
0.0647975355386734,
0.01552598923444748,
-0.014887680299580097,
0.11118700355291367,
0.05733522027730942,
0.022819433361291885,
0.021128227934241295,
0.009885239414870739,
-0.03985453024506569,
-0.013325306586921215,
-0.19498084485530853,
0.06593974679708481,
0.13870294392108917,
0.08520747721195221,
-0.011946371756494045,
0.07486340403556824,
-0.10040875524282455,
-0.12106827646493912,
0.11215335875749588,
-0.06267691403627396,
-0.008283826522529125,
-0.06529007107019424,
0.1326436847448349,
0.14720159769058228,
-0.19087740778923035,
0.06812075525522232,
-0.06286486238241196,
-0.054577119648456573,
-0.11731726676225662,
-0.19296373426914215,
-0.056126803159713745,
-0.0537908673286438,
-0.015372869558632374,
-0.04924733191728592,
0.07244181632995605,
0.056062083691358566,
0.011923623271286488,
0.0036824082490056753,
0.07538694143295288,
-0.01878892257809639,
0.001183385495096445,
0.031768519431352615,
0.06635171920061111,
0.012988004833459854,
-0.030890334397554398,
0.019896110519766808,
-0.008031128905713558,
0.040116652846336365,
0.06276603788137436,
0.03778861090540886,
-0.02647724375128746,
0.016218312084674835,
-0.03651279956102371,
-0.10915597528219223,
0.042008984833955765,
-0.029179582372307777,
-0.08002755790948868,
0.13762354850769043,
0.0252593494951725,
0.005759078077971935,
-0.021546276286244392,
0.24652670323848724,
-0.07283888757228851,
-0.08912359923124313,
-0.14807496964931488,
0.10262012481689453,
-0.03890256956219673,
0.06634368002414703,
0.040788695216178894,
-0.10723847150802612,
0.014206798747181892,
0.12468050420284271,
0.15960600972175598,
-0.04598904773592949,
0.020886680111289024,
0.03171207755804062,
0.0034098350442945957,
-0.03958895057439804,
0.053595054894685745,
0.06899604946374893,
0.1463599056005478,
-0.048738062381744385,
0.08973133563995361,
0.005670548416674137,
-0.10229548066854477,
-0.03761244937777519,
0.11828185617923737,
-0.01828082464635372,
0.019166678190231323,
-0.05744493380188942,
0.11868125200271606,
-0.0608985461294651,
-0.23095959424972534,
0.06514488160610199,
-0.07566448301076889,
-0.14436253905296326,
-0.021428024396300316,
0.07276028394699097,
-0.01780269481241703,
0.026690328493714333,
0.07398071140050888,
-0.07777206599712372,
0.1971588432788849,
0.03926048055291176,
-0.05780704692006111,
-0.05835490673780441,
0.07752813398838043,
-0.1160687804222107,
0.28119364380836487,
0.011715840548276901,
0.05096624419093132,
0.1037345826625824,
-0.015980001538991928,
-0.1366477608680725,
0.01282057911157608,
0.09203975647687912,
-0.08909518271684647,
0.038567181676626205,
0.21223965287208557,
-0.003803485771641135,
0.11778628826141357,
0.07744387537240982,
-0.07308902591466904,
0.04619969055056572,
-0.11599848419427872,
-0.07299992442131042,
-0.08707643300294876,
0.09514311701059341,
-0.07195685058832169,
0.14144541323184967,
0.12355200946331024,
-0.052837129682302475,
0.011311785317957401,
-0.0302458256483078,
0.045323196798563004,
0.00835395511239767,
0.10012959688901901,
0.010363605804741383,
-0.18085741996765137,
0.024026650935411453,
0.015367760322988033,
0.10938689112663269,
-0.17194917798042297,
-0.09750598669052124,
0.0472152978181839,
0.001750203431583941,
-0.0601421594619751,
0.12909571826457977,
0.056909650564193726,
0.04916413500905037,
-0.0436386875808239,
-0.017397185787558556,
-0.009748851880431175,
0.13655178248882294,
-0.1076984852552414,
-0.0029465733096003532
] |
null | null | null | These are some rvc models I trained. I will unpack the game and extract the voice data after training and put it here. | {"language": ["zh"], "license": "mit", "datasets": ["mozi1924/sounds"], "pipeline_tag": "audio-to-audio"} | audio-to-audio | mozi1924/my_rvc_model | [
"audio-to-audio",
"zh",
"dataset:mozi1924/sounds",
"license:mit",
"region:us"
] | 2024-02-06T04:53:12+00:00 | [] | [
"zh"
] | TAGS
#audio-to-audio #zh #dataset-mozi1924/sounds #license-mit #region-us
| These are some rvc models I trained. I will unpack the game and extract the voice data after training and put it here. | [] | [
"TAGS\n#audio-to-audio #zh #dataset-mozi1924/sounds #license-mit #region-us \n"
] | [
32
] | [
"passage: TAGS\n#audio-to-audio #zh #dataset-mozi1924/sounds #license-mit #region-us \n"
] | [
-0.10941426455974579,
-0.03927125781774521,
-0.00287773460149765,
-0.06420425325632095,
0.04559260979294777,
-0.020473584532737732,
0.13743378221988678,
0.05364956706762314,
0.07272553443908691,
0.04222135618329048,
-0.01834135875105858,
0.15503332018852234,
0.037337526679039,
-0.023385891690850258,
0.03137032687664032,
-0.1327863484621048,
0.05375172197818756,
-0.014624547213315964,
0.05661466345191002,
0.01666163094341755,
0.045376017689704895,
-0.024408511817455292,
0.014093903824687004,
0.04295641556382179,
-0.05164523050189018,
0.013275723904371262,
0.01970098167657852,
-0.022552406415343285,
0.0189284048974514,
0.0008770500426180661,
-0.04488430544734001,
0.08415994048118591,
0.012258495204150677,
-0.1859995722770691,
0.026593927294015884,
-0.06186322495341301,
-0.08306553959846497,
-0.03086882084608078,
0.1311461180448532,
-0.09029128402471542,
-0.033478762954473495,
0.06675806641578674,
-0.16587041318416595,
0.08106237649917603,
-0.09316132962703705,
-0.2337949573993683,
-0.10890752822160721,
-0.059080593287944794,
0.009814141318202019,
0.051859404891729355,
-0.02692163549363613,
0.06516872346401215,
-0.14071136713027954,
0.03218547999858856,
0.25264179706573486,
-0.2225756049156189,
0.04541674628853798,
0.08470580726861954,
0.17998473346233368,
0.06952455639839172,
-0.07318737357854843,
0.0989515408873558,
0.04887431859970093,
-0.025740405544638634,
-0.16960255801677704,
-0.0735267847776413,
0.12620976567268372,
0.0654941275715828,
-0.04711555317044258,
0.004803846124559641,
0.3392094373703003,
0.016433371230959892,
-0.04115248844027519,
0.04331706464290619,
0.01063453033566475,
-0.0871620923280716,
-0.02400093525648117,
0.11463480442762375,
-0.008135861717164516,
0.10048015415668488,
-0.05374542623758316,
0.021454837173223495,
-0.09361597895622253,
-0.014416535384953022,
-0.1260722279548645,
0.12954169511795044,
-0.02513381652534008,
0.03773021325469017,
-0.0696641281247139,
-0.0435495562851429,
-0.0021244159433990717,
-0.0736255943775177,
-0.00012399215484037995,
-0.03155410289764404,
-0.019770557060837746,
0.008683562278747559,
-0.005608120933175087,
-0.025439603254199028,
0.17544136941432953,
0.08483195304870605,
0.06811269372701645,
0.0037719777319580317,
-0.07428000867366791,
0.14056400954723358,
0.10970957577228546,
-0.13050878047943115,
0.04349871352314949,
-0.1310635358095169,
-0.02323474921286106,
-0.09531883150339127,
0.06120501458644867,
-0.11186867207288742,
-0.1357080489397049,
-0.00004236400127410889,
-0.0723341777920723,
0.08868507295846939,
0.02585642784833908,
-0.10086117684841156,
-0.1268501579761505,
0.003233829280361533,
-0.0367039293050766,
-0.008740575984120369,
-0.004004080314189196,
0.0006205827812664211,
0.06362656503915787,
0.006366949062794447,
-0.14820252358913422,
0.14481836557388306,
0.04304078221321106,
0.0967772975564003,
-0.06621938198804855,
0.010348107665777206,
0.04009677469730377,
0.050633449107408524,
0.14281204342842102,
0.02425214648246765,
0.10623586177825928,
-0.07639843225479126,
-0.0381174236536026,
-0.030467966571450233,
0.04390235245227814,
0.06675614416599274,
-0.027823396027088165,
0.02708854340016842,
0.03692197799682617,
0.06330844759941101,
-0.12349775433540344,
0.034633636474609375,
-0.11971887201070786,
0.13668426871299744,
-0.05891510471701622,
0.06108484044671059,
-0.14425033330917358,
0.0866960659623146,
-0.07130321860313416,
0.032178401947021484,
0.1363435536623001,
-0.03692914918065071,
-0.12085297703742981,
0.0024159864988178015,
-0.020056793466210365,
-0.03657205402851105,
-0.09771343320608139,
0.08434677869081497,
-0.037048861384391785,
0.053872112184762955,
-0.16832229495048523,
-0.10631795972585678,
0.04404152184724808,
-0.08772280812263489,
-0.057129740715026855,
0.10911261290311813,
0.03526126965880394,
-0.009167782962322235,
0.0662507638335228,
0.4025525450706482,
-0.013358489610254765,
-0.18744878470897675,
0.05530652031302452,
0.06821837276220322,
-0.04267648234963417,
-0.10335791110992432,
0.16366703808307648,
-0.10471762716770172,
-0.07542336732149124,
-0.026248114183545113,
0.16799770295619965,
0.06850789487361908,
-0.08494191616773605,
-0.07667511701583862,
0.027241289615631104,
-0.0506681390106678,
0.05279676616191864,
0.00019233922648709267,
-0.0030983586329966784,
-0.05917617306113243,
0.0596085786819458,
0.019347505643963814,
0.10234024375677109,
0.089151531457901,
-0.002604970009997487,
-0.08347295969724655,
0.18563243746757507,
-0.06169531121850014,
-0.040431343019008636,
-0.09040886163711548,
0.12419484555721283,
-0.05430315062403679,
-0.0031821178272366524,
0.1945684850215912,
0.06202001869678497,
-0.04434267058968544,
-0.031598832458257675,
0.016302568838000298,
-0.020613979548215866,
0.04798181354999542,
0.0556005984544754,
0.047267138957977295,
-0.18456624448299408,
0.049161527305841446,
-0.018343167379498482,
-0.06456795334815979,
0.02014263905584812,
-0.13423438370227814,
0.1684650033712387,
-0.011641542427241802,
-0.027045762166380882,
0.02457480877637863,
0.05608636140823364,
0.005646968260407448,
0.016935601830482483,
0.1061921939253807,
0.0656689926981926,
0.05581829324364662,
0.025986572727560997,
0.22632813453674316,
-0.03659382462501526,
0.2091093510389328,
0.19028879702091217,
-0.08940455317497253,
0.14359335601329803,
-0.06059011444449425,
0.02421533688902855,
-0.06789818406105042,
0.017673274502158165,
-0.07481351494789124,
-0.04064762964844704,
-0.07046432048082352,
0.03583436459302902,
-0.09467877447605133,
0.03462211787700653,
0.05636770278215408,
-0.1296684592962265,
-0.03508184477686882,
0.06565896421670914,
0.03243355453014374,
-0.24038706719875336,
0.1586432158946991,
0.2478247731924057,
0.04992494359612465,
0.43545305728912354,
-0.10815306752920151,
-0.03622741997241974,
-0.06717222183942795,
-0.03215443342924118,
-0.07435975968837738,
0.23239217698574066,
-0.15655159950256348,
-0.007230402901768684,
0.042172156274318695,
0.02929576486349106,
0.0751851499080658,
-0.1586601883172989,
-0.08210146427154541,
-0.04540266841650009,
-0.004556938540190458,
-0.20385925471782684,
0.1167239099740982,
-0.07898535579442978,
-0.016096068546175957,
-0.07610329985618591,
-0.05828580632805824,
0.14381064474582672,
-0.06936828792095184,
-0.11539917439222336,
-0.06793716549873352,
-0.24887681007385254,
-0.05365447327494621,
-0.15496309101581573,
-0.20469054579734802,
-0.07120479643344879,
0.0021546301431953907,
0.1349126100540161,
-0.03904559463262558,
0.021468406543135643,
0.13266155123710632,
0.0437062568962574,
-0.1076614260673523,
-0.06137045845389366,
-0.023643963038921356,
0.04254208132624626,
-0.09440875053405762,
-0.07753831148147583,
-0.02233782596886158,
0.006177463103085756,
0.06890639662742615,
0.06547197699546814,
-0.08557018637657166,
0.09851881861686707,
0.13271503150463104,
0.15037322044372559,
0.021159201860427856,
-0.002320016035810113,
0.1050150990486145,
-0.13014663755893707,
-0.013956649228930473,
0.14038093388080597,
0.013241555541753769,
0.0462927408516407,
0.18204104900360107,
0.09656590968370438,
-0.010112055577337742,
-0.015277470462024212,
-0.009587145410478115,
-0.1135321855545044,
-0.24609464406967163,
-0.15285331010818481,
-0.16438233852386475,
0.009342334233224392,
-0.08017177134752274,
0.04062892496585846,
0.2329990565776825,
-0.04479459300637245,
-0.017429018393158913,
-0.09683326631784439,
-0.009224065579473972,
-0.05042903870344162,
0.02788756974041462,
-0.05212856084108353,
0.015040774829685688,
-0.0824894905090332,
-0.002536331070587039,
0.15182477235794067,
0.09443902969360352,
0.16497890651226044,
0.25552284717559814,
0.13269366323947906,
0.12589801847934723,
0.13573148846626282,
0.15851685404777527,
0.02797742746770382,
0.07356100529432297,
0.007472044788300991,
-0.028751717880368233,
-0.10120290517807007,
0.024949995800852776,
0.053498275578022,
0.23562481999397278,
-0.13643720746040344,
0.000044654640078078955,
-0.14393854141235352,
-0.029061492532491684,
-0.028783075511455536,
0.04469919949769974,
0.020052477717399597,
0.09141431003808975,
0.07817420363426208,
0.07454457879066467,
-0.017714103683829308,
0.1407156139612198,
0.06809720396995544,
-0.03265124186873436,
0.04711693152785301,
0.06789074838161469,
0.03457038104534149,
0.023843247443437576,
0.006178827956318855,
0.014887958765029907,
-0.03848807141184807,
0.013218667358160019,
0.07332155108451843,
-0.13375209271907806,
0.17015856504440308,
0.07767856866121292,
0.0568106584250927,
-0.061243925243616104,
-0.04356050491333008,
0.05075228586792946,
0.01565045863389969,
0.12672008574008942,
0.10242107510566711,
-0.2559141516685486,
-0.05935897305607796,
-0.031946901232004166,
-0.03588083013892174,
0.15034522116184235,
0.07549811899662018,
-0.16925178468227386,
-0.02730759046971798,
-0.009664738550782204,
-0.014456775039434433,
0.08850732445716858,
-0.1459823101758957,
-0.027971386909484863,
0.057979125529527664,
0.16962812840938568,
0.0614926815032959,
-0.08964304625988007,
-0.06968992203474045,
-0.07174055278301239,
-0.06291550397872925,
-0.034747399389743805,
0.033343829214572906,
-0.04409976303577423,
-0.22133970260620117,
0.09003692120313644,
0.03334160894155502,
0.05364585667848587,
0.0023515468928962946,
-0.04595266282558441,
-0.13127020001411438,
-0.10705024003982544,
0.12381444871425629,
-0.07475297898054123,
-0.00029789324617013335,
-0.10197161883115768,
0.27800238132476807,
0.07460379600524902,
0.10172304511070251,
0.01800176501274109,
-0.0044639622792601585,
0.022960178554058075,
-0.060594089329242706,
0.07859721034765244,
-0.1385272741317749,
-0.02412360906600952,
0.03235406428575516,
-0.009877092204988003,
-0.17619198560714722,
0.015446401201188564,
-0.2192302644252777,
0.11588852852582932,
0.19186824560165405,
0.030877824872732162,
0.1937786340713501,
0.3273778557777405,
-0.043266139924526215,
-0.32028043270111084,
-0.12083318829536438,
-0.0980783998966217,
0.02639608271420002,
0.031975168734788895,
-0.17276938259601593,
0.0265944991260767,
0.0033119823783636093,
-0.08512499183416367,
0.22610433399677277,
-0.30994418263435364,
-0.05012349784374237,
0.11276429146528244,
-0.1014367938041687,
0.3668014705181122,
-0.09882483631372452,
-0.16783970594406128,
-0.07348623871803284,
-0.07133639603853226,
0.12948401272296906,
-0.17465537786483765,
0.16029444336891174,
0.1373295783996582,
-0.0009546151850372553,
0.002044557360932231,
-0.011620049364864826,
0.15736271440982819,
0.15299434959888458,
0.007007318083196878,
-0.045373160392045975,
-0.0802592933177948,
0.15179875493049622,
0.07778936624526978,
-0.1320589929819107,
-0.06568294763565063,
-0.019180327653884888,
-0.1336233913898468,
-0.00917220301926136,
-0.08211444318294525,
0.0987294390797615,
0.00797193218022585,
-0.076491579413414,
-0.045431315898895264,
0.07993827015161514,
-0.03481930494308472,
-0.01703392155468464,
0.22630974650382996,
-0.06693996489048004,
-0.06769490242004395,
0.10637368261814117,
-0.04113539680838585,
-0.11988899111747742,
-0.1358744502067566,
-0.07651769369840622,
-0.07819855213165283,
0.08201253414154053,
-0.08271434903144836,
0.05927415192127228,
0.05868585780262947,
0.09820626676082611,
0.0058603351935744286,
0.026125337928533554,
-0.18201519548892975,
0.04900600388646126,
0.14697937667369843,
-0.08975175768136978,
0.02733994461596012,
-0.10737258195877075,
0.14776763319969177,
0.25810331106185913,
-0.028735816478729248,
0.057147666811943054,
0.08916471898555756,
-0.00009868211054708809,
0.012234017252922058,
-0.02768208086490631,
-0.19250324368476868,
0.12034765630960464,
0.04233003407716751,
0.025982314720749855,
-0.17711001634597778,
0.1432952731847763,
-0.04238446056842804,
0.07216343283653259,
-0.06821645051240921,
0.05198763310909271,
-0.028919368982315063,
-0.0995454266667366,
-0.05475278198719025,
0.027297863736748695,
-0.03890226408839226,
-0.12583410739898682,
0.08048512786626816,
-0.08090060204267502,
-0.038141701370477676,
0.09382998198270798,
0.025109194219112396,
0.05670567974448204,
0.0032947356812655926,
-0.06559014320373535,
0.07566305994987488,
-0.021921485662460327,
-0.03735101968050003,
0.07164264470338821,
-0.08048911392688751,
-0.1857936531305313,
-0.010795984417200089,
0.11326755583286285,
-0.08681179583072662,
-0.08575394749641418,
-0.0999065488576889,
0.07236143946647644,
-0.03828703612089157,
-0.023422744125127792,
-0.10260941833257675,
-0.06180582568049431,
0.020679835230112076,
-0.03181168809533119,
-0.03882043808698654,
0.01937931776046753,
-0.07485755532979965,
0.024085935205221176,
0.04316866397857666,
0.12252550572156906,
-0.022335655987262726,
-0.03595726191997528,
0.055707767605781555,
-0.023532766848802567,
0.048109281808137894,
0.17326012253761292,
0.015388194471597672,
0.08227311074733734,
-0.22605641186237335,
-0.1747780442237854,
0.06448327004909515,
0.04440785199403763,
0.002385115483775735,
-0.05629801005125046,
-0.045255232602357864,
0.06369470059871674,
0.00917141605168581,
-0.03968082368373871,
-0.08445826172828674,
-0.10889554023742676,
-0.1048191487789154,
-0.1330144703388214,
-0.10087847709655762,
-0.008153822273015976,
-0.05950238183140755,
0.215516597032547,
0.032019928097724915,
0.12389610707759857,
0.05357813462615013,
0.013309837318956852,
0.0811292827129364,
0.02355145663022995,
-0.038058310747146606,
-0.18123240768909454,
-0.06767280399799347,
-0.08368906378746033,
-0.07260534167289734,
-0.07451734691858292,
0.16114912927150726,
0.003088336903601885,
-0.15493467450141907,
-0.04568697512149811,
0.15864120423793793,
-0.027582990005612373,
0.003776481607928872,
0.3457622528076172,
0.037330400198698044,
0.010255296714603901,
-0.1576904058456421,
-0.028661182150244713,
-0.05472946912050247,
0.09661046415567398,
-0.10666424036026001,
0.08564672619104385,
0.21188779175281525,
-0.02235325425863266,
0.1745632290840149,
-0.09448852390050888,
-0.13361315429210663,
0.05981453135609627,
0.08310846239328384,
-0.02912737987935543,
0.0333283469080925,
0.09124743938446045,
0.15391497313976288,
-0.040044452995061874,
0.06588222831487656,
-0.06029026210308075,
0.008887166157364845,
-0.08256904780864716,
-0.17352540791034698,
-0.01776125282049179,
-0.07912977039813995,
0.043359678238630295,
-0.014015904627740383,
0.12836243212223053,
0.12569621205329895,
0.009589452296495438,
-0.01851712539792061,
0.06806471198797226,
-0.17117299139499664,
-0.17271217703819275,
0.004992765840142965,
-0.07220239192247391,
-0.059503693133592606,
-0.0478183887898922,
-0.09894759207963943,
0.10461602360010147,
-0.12655821442604065,
-0.01718546822667122,
-0.0007476963801309466,
-0.01894371770322323,
0.004233468323945999,
-0.1484006941318512,
-0.06357930600643158,
-0.027904348447918892,
-0.03348095715045929,
0.037847280502319336,
0.18696431815624237,
0.07198059558868408,
-0.019514448940753937,
0.07837972044944763,
0.04303131625056267,
0.029343299567699432,
-0.11338420957326889,
0.06961289048194885,
-0.10247036814689636,
-0.08007964491844177,
0.10643073916435242,
-0.03151332587003708,
0.0016033798456192017,
0.03371245414018631,
0.10503789782524109,
0.2065131813287735,
-0.028785087168216705,
0.0329052209854126,
0.03689158707857132,
0.005741355940699577,
-0.08028313517570496,
0.019094139337539673,
0.02063559554517269,
0.17206835746765137,
-0.008594166487455368,
0.01596829667687416,
-0.17787986993789673,
-0.019178815186023712,
-0.05216681957244873,
0.0751749649643898,
0.09191270917654037,
-0.16538265347480774,
0.0344451479613781,
0.1577664315700531,
-0.23625081777572632,
0.06220530718564987,
0.008757594972848892,
-0.09302625805139542,
-0.02308778092265129,
-0.04876016080379486,
0.10432043671607971,
0.15414348244667053,
0.03258051350712776,
-0.12472249567508698,
-0.061925869435071945,
-0.18667972087860107,
0.04542432352900505,
-0.2475215047597885,
0.0040448615327477455,
-0.02185128815472126,
-0.0007367117796093225,
0.0791458860039711,
-0.030009852722287178,
0.11524488031864166,
-0.007748184259980917,
0.14195939898490906,
0.029398160055279732,
0.15444506704807281,
0.06044813618063927,
-0.1659018099308014,
-0.11044520139694214,
0.11575950682163239,
-0.013054059818387032,
0.06329101324081421,
0.05002545565366745,
0.07398542761802673,
0.07143786549568176,
0.09356837719678879,
-0.03321535512804985,
-0.003264311235398054,
0.05128058046102524,
-0.14775964617729187,
0.11671948432922363,
-0.1942654699087143,
0.02952565811574459,
-0.09282728284597397,
-0.014487873762845993,
0.01573329232633114,
0.12381890416145325,
-0.08275768160820007,
0.02107989974319935,
-0.018340252339839935,
-0.013228977099061012,
-0.0010605736169964075,
0.0030002882704138756,
-0.11722084134817123,
-0.00428770249709487,
-0.09920377284288406,
0.0839354395866394,
-0.12699826061725616,
0.018116921186447144,
0.1024492084980011,
-0.06695601344108582,
0.03468092158436775,
-0.22989174723625183,
0.055962640792131424,
-0.031190570443868637,
-0.07304208725690842,
-0.07528534531593323
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# test_1_3136_files
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the audiofolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2746
- Accuracy: 0.9331
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 20
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.6934 | 0.96 | 19 | 0.6904 | 0.5334 |
| 0.6794 | 1.97 | 39 | 0.6392 | 0.6672 |
| 0.5873 | 2.99 | 59 | 0.5179 | 0.7548 |
| 0.4796 | 4.0 | 79 | 0.3624 | 0.8710 |
| 0.3158 | 4.96 | 98 | 0.3010 | 0.8933 |
| 0.2043 | 5.97 | 118 | 0.5355 | 0.8201 |
| 0.1699 | 6.99 | 138 | 0.2456 | 0.9220 |
| 0.1443 | 8.0 | 158 | 0.2069 | 0.9268 |
| 0.1258 | 8.96 | 177 | 0.2480 | 0.9172 |
| 0.1057 | 9.97 | 197 | 0.2855 | 0.9188 |
| 0.0796 | 10.99 | 217 | 0.3025 | 0.9108 |
| 0.0582 | 12.0 | 237 | 0.3201 | 0.9061 |
| 0.0505 | 12.96 | 256 | 0.2977 | 0.9188 |
| 0.0802 | 13.97 | 276 | 0.1667 | 0.9490 |
| 0.0377 | 14.99 | 296 | 0.2239 | 0.9411 |
| 0.0346 | 16.0 | 316 | 0.2396 | 0.9315 |
| 0.0372 | 16.96 | 335 | 0.2970 | 0.9204 |
| 0.0222 | 17.97 | 355 | 0.3063 | 0.9204 |
| 0.0214 | 18.99 | 375 | 0.2761 | 0.9331 |
| 0.0192 | 19.24 | 380 | 0.2746 | 0.9331 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["audiofolder"], "metrics": ["accuracy"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "test_1_3136_files", "results": [{"task": {"type": "audio-classification", "name": "Audio Classification"}, "dataset": {"name": "audiofolder", "type": "audiofolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9331210191082803, "name": "Accuracy"}]}]}]} | audio-classification | PatricioMN/test_1_3136_files | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"audio-classification",
"generated_from_trainer",
"dataset:audiofolder",
"base_model:facebook/wav2vec2-base",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-06T04:54:37+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us
| test\_1\_3136\_files
====================
This model is a fine-tuned version of facebook/wav2vec2-base on the audiofolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2746
* Accuracy: 0.9331
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 20
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 20",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 20",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
78,
144,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 20### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.14518699049949646,
0.1497461497783661,
-0.0016317948466166854,
0.06461835652589798,
0.12669715285301208,
0.00823906995356083,
0.11022841185331345,
0.11835051327943802,
-0.09816300868988037,
0.11248230189085007,
0.10388914495706558,
0.08693575114011765,
0.053181812167167664,
0.14719446003437042,
-0.03364459425210953,
-0.2824122905731201,
0.008604765869677067,
0.0019560521468520164,
-0.15648892521858215,
0.11685680598020554,
0.07965056598186493,
-0.1041334941983223,
0.060608766973018646,
0.014018118381500244,
-0.1495579481124878,
-0.0037794935051351786,
-0.022214243188500404,
-0.06428966671228409,
0.09046266973018646,
0.042433176189661026,
0.0839700922369957,
0.04885650426149368,
0.09077342599630356,
-0.2142784744501114,
0.014838410541415215,
0.08421894162893295,
0.00403218949213624,
0.09056759625673294,
0.10517764836549759,
-0.022703861817717552,
0.09995236247777939,
-0.07261347025632858,
0.05911490321159363,
0.050479236990213394,
-0.09571369737386703,
-0.2950039207935333,
-0.09806367009878159,
0.07857980579137802,
0.12163381278514862,
0.08643219619989395,
-0.02994043380022049,
0.06399768590927124,
-0.05106528848409653,
0.09354987740516663,
0.24914340674877167,
-0.2617466151714325,
-0.07318137586116791,
0.041267551481723785,
0.07398245483636856,
0.061303865164518356,
-0.1182493343949318,
-0.0031148463021963835,
0.06286590546369553,
0.020287465304136276,
0.10695755481719971,
0.004179023206233978,
0.05585348606109619,
-0.01041088905185461,
-0.14969177544116974,
-0.03660086169838905,
0.1607970893383026,
0.10711584985256195,
-0.04543279856443405,
-0.07157505303621292,
-0.0324152372777462,
-0.22827878594398499,
-0.03392050415277481,
-0.001748628681525588,
0.03342190757393837,
-0.052223093807697296,
-0.12122073769569397,
0.029384374618530273,
-0.06842728704214096,
-0.1011442244052887,
0.0377577506005764,
0.12295348197221756,
0.04133808612823486,
-0.019019443541765213,
0.008884071372449398,
0.11468183249235153,
0.036612335592508316,
-0.15963415801525116,
0.0015947931678965688,
0.024227768182754517,
-0.09174910187721252,
-0.024788569658994675,
-0.022378234192728996,
-0.022181084379553795,
0.004533455241471529,
0.153030663728714,
-0.06940743327140808,
0.0628470703959465,
0.03849957510828972,
0.04258839040994644,
-0.06686132401227951,
0.12251550704240799,
-0.08984409272670746,
-0.09711190313100815,
-0.0401524119079113,
0.1029675230383873,
0.009178156033158302,
-0.022782539948821068,
-0.0885557234287262,
0.04423386603593826,
0.10190732777118683,
0.019218187779188156,
-0.018982356414198875,
0.017616016790270805,
-0.07226841151714325,
-0.03353647142648697,
0.04546664282679558,
-0.07797997444868088,
0.038593363016843796,
0.029687024652957916,
-0.06147884577512741,
-0.016496097669005394,
0.01620425283908844,
0.026528144255280495,
0.028515085577964783,
0.14838047325611115,
-0.09198471903800964,
-0.03778121620416641,
-0.06995493173599243,
-0.08149892091751099,
0.0418219156563282,
-0.07410489767789841,
0.019471149891614914,
-0.06270735710859299,
-0.12069299072027206,
-0.04874483495950699,
0.07606112957000732,
-0.04064745828509331,
-0.08760130405426025,
-0.04800767824053764,
-0.10411231964826584,
0.050454024225473404,
-0.024700403213500977,
0.14182960987091064,
-0.06318934261798859,
0.10949937254190445,
0.01730521209537983,
0.06023554503917694,
0.038483649492263794,
0.06949114799499512,
-0.04047837108373642,
0.06886404752731323,
-0.1913910210132599,
0.051388803869485855,
-0.10521747171878815,
0.03965466842055321,
-0.12112797051668167,
-0.10752642899751663,
-0.031664807349443436,
0.01649658940732479,
0.06469887495040894,
0.08577518910169601,
-0.1840229034423828,
-0.12231723219156265,
0.1707434058189392,
-0.08441456407308578,
-0.11778111755847931,
0.11716021597385406,
-0.008911136537790298,
-0.026390312239527702,
0.03484533727169037,
0.1531207263469696,
0.11189841479063034,
-0.1142057403922081,
-0.04388093948364258,
-0.030861809849739075,
0.09492619335651398,
0.001493224990554154,
0.113525390625,
-0.02735232375562191,
0.03317144140601158,
-0.010260498151183128,
-0.033277031034231186,
0.04300404712557793,
-0.09257975965738297,
-0.07217737287282944,
-0.017749793827533722,
-0.09467536956071854,
0.024739282205700874,
0.055807530879974365,
0.025423642247915268,
-0.07991012930870056,
-0.14111949503421783,
0.05408449098467827,
0.11845690757036209,
-0.07967871427536011,
0.006225490942597389,
-0.06408799439668655,
0.09633879363536835,
-0.07007541507482529,
-0.029430408030748367,
-0.15687443315982819,
-0.0260902252048254,
0.014541865326464176,
-0.048605699092149734,
0.004260867368429899,
0.016488080844283104,
0.06363696604967117,
0.08061801642179489,
-0.05941389128565788,
-0.09159240126609802,
-0.06060777232050896,
0.003468484617769718,
-0.06949848681688309,
-0.2587120532989502,
-0.07428088784217834,
-0.028585070744156837,
0.1587112694978714,
-0.24320603907108307,
0.009662219323217869,
0.03588084876537323,
0.1394215226173401,
0.06035258620977402,
-0.05628853291273117,
-0.0024205849040299654,
0.05693160742521286,
-0.02252989076077938,
-0.07404518127441406,
0.020475557073950768,
0.009942936711013317,
-0.09670835733413696,
-0.02544490434229374,
-0.12120970338582993,
0.1583007276058197,
0.10711585730314255,
0.014815299771726131,
-0.09141966700553894,
-0.026358384639024734,
-0.08672036230564117,
-0.056236766278743744,
-0.029159320518374443,
-0.007555366959422827,
0.11667576432228088,
0.027411021292209625,
0.11999735236167908,
-0.0885629802942276,
-0.048437390476465225,
0.05697259679436684,
-0.003249007510021329,
-0.02297765202820301,
0.12704353034496307,
0.08389198035001755,
-0.07148580998182297,
0.14531944692134857,
0.12219860404729843,
-0.05855831503868103,
0.18037956953048706,
-0.06705859303474426,
-0.11422401666641235,
-0.027818778529763222,
0.015322558581829071,
0.02380373701453209,
0.1430160254240036,
-0.10878288745880127,
0.00913725420832634,
0.01648627035319805,
0.03325214236974716,
0.0160992331802845,
-0.18111109733581543,
-0.016610311344265938,
0.04625649005174637,
-0.053269967436790466,
-0.03730876371264458,
-0.005185484886169434,
-0.01848326437175274,
0.07529208809137344,
0.006274252664297819,
-0.05204841494560242,
0.013411708176136017,
0.0008353231241926551,
-0.07899820059537888,
0.18523533642292023,
-0.09367936104536057,
-0.13127978146076202,
-0.15681928396224976,
-0.017742514610290527,
-0.03246217221021652,
-0.005355077795684338,
0.05641213804483414,
-0.09442620724439621,
-0.032549381256103516,
-0.050656117498874664,
0.04541494697332382,
-0.028511881828308105,
0.03298751637339592,
0.030385300517082214,
0.02303425781428814,
0.09124062210321426,
-0.10041876882314682,
0.03655070438981056,
-0.007574568502604961,
-0.03133295476436615,
0.001409623771905899,
0.038605447858572006,
0.09289737045764923,
0.16205041110515594,
0.05055150017142296,
0.02190444990992546,
-0.021498247981071472,
0.19574861228466034,
-0.12912429869174957,
0.0036234171129763126,
0.10532740503549576,
-0.0190704558044672,
0.046133577823638916,
0.14221617579460144,
0.0546036995947361,
-0.0823746845126152,
0.018557023257017136,
0.07712578773498535,
-0.030685750767588615,
-0.23622362315654755,
-0.021253937855362892,
-0.04269850626587868,
0.02199683152139187,
0.08703488856554031,
0.036325447261333466,
0.040439486503601074,
0.052758313715457916,
-0.015062352642416954,
0.031275928020477295,
-0.01068643294274807,
0.05396635830402374,
0.0012335714418441057,
0.037231504917144775,
0.11907613277435303,
-0.024945583194494247,
-0.00733970245346427,
0.03911386430263519,
0.009115387685596943,
0.2314188927412033,
0.00465830322355032,
0.14159445464611053,
0.07718013972043991,
0.136166512966156,
0.015495902858674526,
0.058958277106285095,
-0.00006078109436202794,
-0.034953370690345764,
0.014069548808038235,
-0.057581909000873566,
0.000008106630957627203,
0.04676864668726921,
0.03425474837422371,
0.056501127779483795,
-0.13125906884670258,
0.040084730833768845,
0.014206994324922562,
0.30990004539489746,
0.07490907609462738,
-0.32077255845069885,
-0.10865360498428345,
0.008344984613358974,
-0.05103002488613129,
-0.048597969114780426,
0.03462270274758339,
0.1440567523241043,
-0.0777197778224945,
0.07481572031974792,
-0.07018282264471054,
0.08885209262371063,
-0.04949087277054787,
-0.0033437670208513737,
0.10932867228984833,
0.10639578849077225,
-0.017538627609610558,
0.0601612851023674,
-0.20229482650756836,
0.28620657324790955,
0.011143062263727188,
0.07061250507831573,
-0.026025036349892616,
0.03024384379386902,
0.025192802771925926,
0.02398848719894886,
0.11322066187858582,
-0.009557339362800121,
-0.10281699150800705,
-0.17192360758781433,
-0.099723219871521,
-0.0023816090542823076,
0.12361754477024078,
-0.09086479246616364,
0.10212709754705429,
-0.022715575993061066,
-0.03540588542819023,
0.05880401283502579,
-0.04859517887234688,
-0.10080204159021378,
-0.10252454876899719,
0.01152514573186636,
0.003651238512247801,
0.0487547293305397,
-0.10686600208282471,
-0.11752743273973465,
-0.11158271133899689,
0.14556007087230682,
-0.10431531816720963,
-0.015031101182103157,
-0.13574005663394928,
0.08684387803077698,
0.14366821944713593,
-0.06707780063152313,
0.06623492389917374,
0.01795956864953041,
0.14395995438098907,
0.02018459513783455,
-0.02988823875784874,
0.11448453366756439,
-0.09788374602794647,
-0.22682657837867737,
-0.06437873095273972,
0.17436739802360535,
0.05773060768842697,
0.053781602531671524,
-0.037087634205818176,
0.027468210086226463,
0.014238344505429268,
-0.07229741662740707,
0.07671809196472168,
-0.010616859421133995,
0.04714386165142059,
0.01718779094517231,
-0.013760543428361416,
0.0030393782071769238,
-0.022271789610385895,
-0.05242343246936798,
0.09757260978221893,
0.2909053564071655,
-0.08684193342924118,
0.03619108349084854,
0.05763598904013634,
-0.03734660893678665,
-0.15206173062324524,
0.06019984558224678,
0.14193490147590637,
0.030119193717837334,
0.03945273533463478,
-0.206192746758461,
0.10411710292100906,
0.09062547236680984,
-0.030590975657105446,
0.1227073222398758,
-0.2798035144805908,
-0.11843734234571457,
0.09768527001142502,
0.09497471153736115,
-0.03444583714008331,
-0.1580740511417389,
-0.0734669417142868,
-0.013468857854604721,
-0.14304637908935547,
0.09071644395589828,
-0.09801512956619263,
0.1181788519024849,
0.010177407413721085,
0.018863623961806297,
0.010147119872272015,
-0.051585085690021515,
0.12362933903932571,
0.020935047417879105,
0.07388004660606384,
-0.00884193368256092,
0.026054563000798225,
0.02738909423351288,
-0.06108151003718376,
-0.009431174024939537,
-0.0920848473906517,
0.026913149282336235,
-0.09112131595611572,
-0.02824913151562214,
-0.0816526710987091,
0.033315736800432205,
-0.06436285376548767,
-0.042502786964178085,
-0.025016114115715027,
0.054946210235357285,
0.04663017392158508,
-0.017302371561527252,
0.1442229151725769,
-0.012746877036988735,
0.14613406360149384,
0.12234723567962646,
0.06779422610998154,
-0.008934758603572845,
-0.12059983611106873,
-0.0186585895717144,
-0.028947409242391586,
0.06305444985628128,
-0.13605345785617828,
0.030899861827492714,
0.12062420696020126,
0.048349857330322266,
0.13588926196098328,
0.054427750408649445,
-0.09082715213298798,
0.0015358903910964727,
0.08347341418266296,
-0.10260765254497528,
-0.110537588596344,
-0.03296295925974846,
-0.009781736880540848,
-0.13075077533721924,
0.04110565409064293,
0.1011565551161766,
-0.042883165180683136,
0.005459410138428211,
0.013047645799815655,
0.022022413089871407,
-0.05863416567444801,
0.20930595695972443,
0.05601995065808296,
0.08700675517320633,
-0.09829563647508621,
0.08799506723880768,
0.03200310468673706,
-0.14246965944766998,
0.029426805675029755,
0.05465352535247803,
-0.059396859258413315,
-0.004502111580222845,
0.03373429551720619,
0.11304258555173874,
-0.0004591978504322469,
-0.06513450294733047,
-0.1135491207242012,
-0.12476418912410736,
0.06902850419282913,
0.12909668684005737,
0.043798819184303284,
0.02905372716486454,
-0.021049782633781433,
0.05486011132597923,
-0.11291753500699997,
0.12986703217029572,
0.08025944977998734,
0.09142553061246872,
-0.2085198163986206,
0.10683660954236984,
0.01185187604278326,
0.0008192053064703941,
-0.010736025869846344,
0.02918541617691517,
-0.11716301739215851,
-0.002346306573599577,
-0.10343561321496964,
-0.03757983446121216,
-0.05580972507596016,
-0.008233267813920975,
-0.00609935587272048,
-0.04673406481742859,
-0.06640621274709702,
0.024634018540382385,
-0.10892413556575775,
-0.041772596538066864,
0.012065060436725616,
0.06703239679336548,
-0.1167992353439331,
0.0007112182211130857,
0.039522167295217514,
-0.10753484070301056,
0.08035976439714432,
0.038831524550914764,
0.04690805450081825,
0.021376097574830055,
-0.1180301383137703,
0.01829000748693943,
0.04879794269800186,
-0.01821589469909668,
0.053747065365314484,
-0.15924760699272156,
-0.011849889531731606,
-0.04865940660238266,
0.019971689209342003,
-0.012773416936397552,
0.021032603457570076,
-0.12383926659822464,
-0.025339998304843903,
-0.04600651562213898,
-0.03961820527911186,
-0.056976567953825,
0.043893903493881226,
0.09876564890146255,
0.011801247484982014,
0.18497610092163086,
-0.06788291037082672,
0.028506508097052574,
-0.227646604180336,
0.008630380034446716,
-0.008816019631922245,
-0.0789889320731163,
-0.059474844485521317,
-0.01354892272502184,
0.06761632114648819,
-0.0661574974656105,
0.0873008593916893,
-0.054740432649850845,
0.02981017716228962,
0.025217309594154358,
-0.09605779498815536,
0.03825908899307251,
0.04891060292720795,
0.1859952211380005,
0.0266630370169878,
-0.011191044934093952,
0.04109715670347214,
0.019096575677394867,
0.08045149594545364,
0.06824691593647003,
0.16194655001163483,
0.14330247044563293,
-0.03623406961560249,
0.076656773686409,
0.06937827914953232,
-0.12978997826576233,
-0.1807837188243866,
0.11752971261739731,
-0.06447427719831467,
0.1311471313238144,
0.0053785271011292934,
0.17134538292884827,
0.11043663322925568,
-0.20083943009376526,
0.04002914950251579,
-0.030527057126164436,
-0.07844901084899902,
-0.10342931747436523,
-0.06709770113229752,
-0.07791630178689957,
-0.19996851682662964,
0.02641880325973034,
-0.114165298640728,
0.03530697152018547,
0.04982319101691246,
0.02469458058476448,
0.0302130077034235,
0.17573808133602142,
0.02243388630449772,
0.020782867446541786,
0.07359365373849869,
0.02842565067112446,
-0.05212416127324104,
-0.04728463292121887,
-0.08345063775777817,
0.05102000758051872,
-0.039021968841552734,
0.0371846966445446,
-0.08100085705518723,
-0.09741171449422836,
0.08708266913890839,
0.04398176819086075,
-0.09791041910648346,
0.02488509565591812,
-0.006284022703766823,
0.05155106633901596,
0.06230158358812332,
0.005332578904926777,
0.006994296796619892,
-0.020237121731042862,
0.22143377363681793,
-0.09977179765701294,
-0.002577833831310272,
-0.1440906971693039,
0.1831008791923523,
-0.011691291816532612,
-0.0015664532547816634,
0.028354214504361153,
-0.09768502414226532,
0.0014711314579471946,
0.15664619207382202,
0.1525813490152359,
-0.013104473240673542,
-0.023571087047457695,
0.024007944390177727,
-0.014683689922094345,
-0.062377914786338806,
0.05884505808353424,
0.11977704614400864,
0.09612466394901276,
-0.06026658043265343,
-0.05742601305246353,
-0.04891116917133331,
-0.05646370351314545,
0.016429198905825615,
0.09034176915884018,
0.03705713152885437,
-0.019358085468411446,
-0.015276224352419376,
0.10812562704086304,
-0.0737885981798172,
-0.16813817620277405,
0.049409639090299606,
-0.16766488552093506,
-0.1849396973848343,
-0.058619190007448196,
0.08698997646570206,
0.018270011991262436,
0.050252314656972885,
0.0008993836818262935,
-0.046745702624320984,
0.10413743555545807,
0.000721031567081809,
-0.051547612994909286,
-0.11095026880502701,
0.052431121468544006,
-0.08301853388547897,
0.20591624081134796,
-0.05657326802611351,
0.021002812311053276,
0.11347884684801102,
0.055897071957588196,
-0.07104618102312088,
0.02248738519847393,
0.0676017552614212,
-0.1322820484638214,
0.017813200131058693,
0.1954377293586731,
-0.04736487194895744,
0.1495889574289322,
0.03527704253792763,
-0.13597069680690765,
0.006796761881560087,
-0.09137818962335587,
-0.06766516715288162,
-0.07181418687105179,
-0.006021153647452593,
-0.05182328075170517,
0.13916973769664764,
0.1963227540254593,
-0.05756956711411476,
-0.01675909385085106,
-0.056331295520067215,
0.04727235436439514,
0.07252872735261917,
0.11059845238924026,
0.005179709289222956,
-0.27659836411476135,
0.04097919538617134,
0.009229893796145916,
-0.0007753216195851564,
-0.2560436725616455,
-0.08008289337158203,
0.023385625332593918,
-0.06189814582467079,
-0.07302529364824295,
0.07745256274938583,
0.07094784080982208,
0.05751288682222366,
-0.06780523806810379,
-0.022904694080352783,
-0.06433968245983124,
0.17054660618305206,
-0.1817556470632553,
-0.07058537006378174
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-cased-lora-592K-snli-model2
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.9156
- Accuracy: 0.6375
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 256
- eval_batch_size: 256
- seed: 98
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.6215 | 1.0 | 2146 | 0.5300 | 0.7996 |
| 0.5619 | 2.0 | 4292 | 0.4853 | 0.8164 |
| 0.5446 | 3.0 | 6438 | 0.4782 | 0.8188 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-base-cased-lora-592K-snli-model2", "results": []}]} | text-classification | varun-v-rao/bert-base-cased-lora-592K-snli-model2 | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:bert-base-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T04:58:03+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-base-cased-lora-592K-snli-model2
=====================================
This model is a fine-tuned version of bert-base-cased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.9156
* Accuracy: 0.6375
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 256
* eval\_batch\_size: 256
* seed: 98
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 98\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 98\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
67,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 98\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.08894699811935425,
0.108758844435215,
-0.002147892490029335,
0.10633673518896103,
0.1493767946958542,
0.021484412252902985,
0.15700002014636993,
0.10932836681604385,
-0.07722398638725281,
0.04654308035969734,
0.1250632107257843,
0.12847165763378143,
0.015215017832815647,
0.1305249035358429,
-0.061375945806503296,
-0.22570475935935974,
0.009530636481940746,
0.03258287161588669,
-0.06905625760555267,
0.1170974150300026,
0.09107226133346558,
-0.13025516271591187,
0.09148183465003967,
-0.009747802279889584,
-0.17933256924152374,
0.019446738064289093,
0.021819939836859703,
-0.05271757394075394,
0.1359519362449646,
0.03446013480424881,
0.13936284184455872,
0.020732171833515167,
0.09175645560026169,
-0.21160490810871124,
0.011289632879197598,
0.06426176428794861,
-0.006358297076076269,
0.08251260221004486,
0.0335937924683094,
0.010634936392307281,
0.09162142872810364,
-0.08021347969770432,
0.06439249217510223,
0.017769664525985718,
-0.11812204122543335,
-0.19420388340950012,
-0.07377538830041885,
0.04333006963133812,
0.09373460710048676,
0.0777018740773201,
-0.011047469452023506,
0.12266390025615692,
-0.05471616983413696,
0.08928356319665909,
0.209172785282135,
-0.3218810260295868,
-0.06570690125226974,
0.05104948580265045,
0.029024992138147354,
0.08572413772344589,
-0.10461760312318802,
-0.01754320226609707,
0.06827011704444885,
0.027878666296601295,
0.13128887116909027,
-0.028147229924798012,
-0.07424253970384598,
0.008127332665026188,
-0.14947567880153656,
-0.017314793542027473,
0.14808815717697144,
0.050826605409383774,
-0.041590675711631775,
-0.04480651393532753,
-0.06325623393058777,
-0.15245002508163452,
-0.04022888466715813,
-0.02872336655855179,
0.05144287645816803,
-0.01989242620766163,
-0.05915794149041176,
-0.010199167765676975,
-0.1078413724899292,
-0.08093104511499405,
-0.05712512135505676,
0.1339530497789383,
0.04024043679237366,
0.0011648127110674977,
-0.012532302178442478,
0.09795620292425156,
-0.033323246985673904,
-0.11955665051937103,
0.02380378730595112,
0.022631758823990822,
0.010137856006622314,
-0.05667309835553169,
-0.053880419582128525,
-0.06121465191245079,
0.028104934841394424,
0.13240957260131836,
-0.05864013358950615,
0.05418051406741142,
0.009730123914778233,
0.04630984365940094,
-0.10336639732122421,
0.1574493646621704,
-0.039817556738853455,
-0.025517035275697708,
0.01571931689977646,
0.07994119077920914,
0.04125028848648071,
-0.0013802021276205778,
-0.127062126994133,
0.030864551663398743,
0.11080887168645859,
0.016492314636707306,
-0.08131926506757736,
0.07983002066612244,
-0.05053005367517471,
0.0028030527755618095,
0.016713589429855347,
-0.08791054040193558,
0.030010424554347992,
0.007396161090582609,
-0.051533907651901245,
-0.07599706202745438,
0.029972776770591736,
0.018438654020428658,
0.0037520593032240868,
0.10165636986494064,
-0.08695709705352783,
0.00805208645761013,
-0.09424316138029099,
-0.12312320619821548,
0.02110162377357483,
-0.07773274183273315,
0.02139618806540966,
-0.11090442538261414,
-0.15591956675052643,
-0.008374301716685295,
0.057493481785058975,
-0.023693202063441277,
-0.030017638579010963,
-0.056181684136390686,
-0.07326342910528183,
0.012826595455408096,
-0.018147585913538933,
0.0929442048072815,
-0.06603901833295822,
0.09298735111951828,
0.0355554036796093,
0.06604073196649551,
-0.047805823385715485,
0.03732560947537422,
-0.09763193875551224,
0.02615414559841156,
-0.17877870798110962,
0.011979036033153534,
-0.06941812485456467,
0.05830948054790497,
-0.08282909542322159,
-0.07493732124567032,
-0.011575642973184586,
0.01681828871369362,
0.0727275013923645,
0.09093345701694489,
-0.16588298976421356,
-0.0628867894411087,
0.17716172337532043,
-0.0970221757888794,
-0.1415255069732666,
0.12315904349088669,
-0.05980651080608368,
0.05261177197098732,
0.06646409630775452,
0.18114371597766876,
0.06458105146884918,
-0.09791085124015808,
0.0028114020824432373,
0.0015697096241638064,
0.06279904395341873,
-0.05098661407828331,
0.06119269132614136,
0.0004865051887463778,
-0.014049041084945202,
0.016958532854914665,
-0.04782755300402641,
0.050031986087560654,
-0.07998383045196533,
-0.08472544699907303,
-0.037816453725099564,
-0.10216444730758667,
0.04587322100996971,
0.05776060000061989,
0.06261298805475235,
-0.11635556817054749,
-0.09029926359653473,
0.0724758431315422,
0.07274043560028076,
-0.07359109073877335,
0.018050838261842728,
-0.07269325107336044,
0.0775098130106926,
-0.055269546806812286,
-0.01792897842824459,
-0.15123014152050018,
-0.04826835170388222,
0.01577710546553135,
0.0000014030831607669825,
0.011913580819964409,
0.007336189970374107,
0.07176210731267929,
0.0760161504149437,
-0.06747224926948547,
-0.02530038170516491,
-0.015098460018634796,
0.019640052691102028,
-0.13003619015216827,
-0.2074243575334549,
-0.015679555013775826,
-0.032786935567855835,
0.13704101741313934,
-0.24339282512664795,
0.052726708352565765,
0.0017806467367336154,
0.07915288209915161,
0.03438648581504822,
-0.004248298704624176,
-0.046406783163547516,
0.07695866376161575,
-0.042382095009088516,
-0.05441097542643547,
0.06138807162642479,
0.013120847754180431,
-0.08171174675226212,
-0.03878093138337135,
-0.14665794372558594,
0.19055959582328796,
0.13393162190914154,
-0.09565512090921402,
-0.0849575400352478,
-0.014034666121006012,
-0.042242638766765594,
-0.029159490019083023,
-0.04639464616775513,
0.004261899273842573,
0.1337875872850418,
-0.018441954627633095,
0.15023860335350037,
-0.08027993887662888,
-0.03191956877708435,
0.024068554863333702,
-0.04701456055045128,
0.009305551648139954,
0.11117786169052124,
0.11843337118625641,
-0.10839078575372696,
0.15457966923713684,
0.1738593578338623,
-0.0985608845949173,
0.13615384697914124,
-0.04199287295341492,
-0.056395214051008224,
-0.02100224420428276,
0.007272660266608,
0.004818096291273832,
0.1200045570731163,
-0.13282285630702972,
-0.004549501463770866,
0.0025104752276092768,
0.013895955868065357,
0.01913316361606121,
-0.22591355443000793,
-0.02721198834478855,
0.03756193816661835,
-0.049629949033260345,
0.012697417289018631,
-0.03114701807498932,
-0.012252485379576683,
0.10062456876039505,
-0.002031970303505659,
-0.08813624083995819,
0.042649175971746445,
-0.005483677145093679,
-0.0911484882235527,
0.21331022679805756,
-0.08384082466363907,
-0.13618795573711395,
-0.132440984249115,
-0.06400319188833237,
-0.03331984207034111,
0.025712206959724426,
0.0622527040541172,
-0.08078977465629578,
-0.038634203374385834,
-0.10928058624267578,
-0.0027101566083729267,
0.03365623205900192,
0.029385671019554138,
0.009765872731804848,
0.004675242584198713,
0.0825572982430458,
-0.10219328850507736,
-0.012174549512565136,
-0.05213196575641632,
-0.0609046071767807,
0.029110150411725044,
0.02619868703186512,
0.1158333495259285,
0.15029732882976532,
-0.03201783820986748,
-0.002104603685438633,
-0.034508056938648224,
0.21502716839313507,
-0.05421394482254982,
-0.017691755667328835,
0.1185658872127533,
-0.028712501749396324,
0.04349305480718613,
0.1368733048439026,
0.06444772332906723,
-0.09388954192399979,
0.02513801120221615,
0.044114917516708374,
-0.028746409341692924,
-0.214349165558815,
-0.03585448116064072,
-0.033957939594984055,
0.00001291054104513023,
0.09719280898571014,
0.03564545884728432,
0.02424550987780094,
0.06837224215269089,
0.03602907806634903,
0.07266412675380707,
-0.01359912846237421,
0.06527602672576904,
0.1145465075969696,
0.03829514607787132,
0.12596438825130463,
-0.05205615237355232,
-0.06140035763382912,
0.03233847767114639,
0.00311234500259161,
0.20120057463645935,
0.028880679979920387,
0.12440600991249084,
0.06387592107057571,
0.1595788598060608,
0.0026390820275992155,
0.06771726906299591,
-0.012330234050750732,
-0.04815477877855301,
-0.010434874333441257,
-0.050591107457876205,
-0.03260234370827675,
0.039238158613443375,
-0.09681446105241776,
0.05911022052168846,
-0.11521057039499283,
0.016773933544754982,
0.058905743062496185,
0.23703761398792267,
0.0447111651301384,
-0.31975600123405457,
-0.08647389709949493,
0.023708034306764603,
-0.026138462126255035,
-0.0228952094912529,
0.0365692675113678,
0.1368900090456009,
-0.049750298261642456,
0.02312873862683773,
-0.06968153268098831,
0.07776302099227905,
-0.04374045133590698,
0.04420221596956253,
0.06698859483003616,
0.08542072772979736,
-0.009148851968348026,
0.07068303227424622,
-0.2806471288204193,
0.27413079142570496,
0.012582342140376568,
0.07507665455341339,
-0.0516047328710556,
0.0029533400665968657,
0.03254235163331032,
0.08960070461034775,
0.07491574436426163,
-0.022756410762667656,
-0.07606224715709686,
-0.20009459555149078,
-0.052133943885564804,
0.031808920204639435,
0.0933651477098465,
-0.029812108725309372,
0.09671413153409958,
-0.03416481986641884,
0.0038808370009064674,
0.09188546240329742,
-0.014403948560357094,
-0.08530979603528976,
-0.09879347681999207,
-0.025665318593382835,
0.039823148399591446,
-0.03201349824666977,
-0.08475508540868759,
-0.10460900515317917,
-0.13372430205345154,
0.15534597635269165,
-0.05923164263367653,
-0.021224703639745712,
-0.09298361837863922,
0.06701824814081192,
0.040988989174366,
-0.07489900290966034,
0.050755489617586136,
0.010308324359357357,
0.08725108206272125,
0.02394801937043667,
-0.05811883881688118,
0.12877164781093597,
-0.0757623240351677,
-0.16530705988407135,
-0.0782139003276825,
0.09958585351705551,
0.027257762849330902,
0.04545203223824501,
0.0006868485943414271,
0.003235210431739688,
-0.006868899334222078,
-0.07792002707719803,
0.022982247173786163,
-0.0039227185770869255,
0.0657384917140007,
0.006412850692868233,
-0.08756009489297867,
-0.004813432227820158,
-0.05158830061554909,
-0.030160902068018913,
0.16511426866054535,
0.2661231458187103,
-0.09263218194246292,
-0.0009609082480892539,
0.06670571118593216,
-0.07611387223005295,
-0.21537715196609497,
0.04669758677482605,
0.028826914727687836,
0.003128761425614357,
0.043310195207595825,
-0.15376269817352295,
0.12223675101995468,
0.10198145359754562,
-0.02809414267539978,
0.10361658036708832,
-0.28530174493789673,
-0.13121207058429718,
0.13951675593852997,
0.16672471165657043,
0.10656418651342392,
-0.16055378317832947,
-0.030992038547992706,
-0.032729316502809525,
-0.10315726697444916,
0.11426709592342377,
-0.12866926193237305,
0.10427988320589066,
-0.004080511163920164,
0.050542764365673065,
0.0008757206378504634,
-0.05711245909333229,
0.11897104233503342,
-0.01748049445450306,
0.1138589084148407,
-0.06206327676773071,
-0.023715635761618614,
0.045515697449445724,
-0.05338608846068382,
0.01859080046415329,
-0.1168021410703659,
0.030020643025636673,
-0.053096551448106766,
-0.031742826104164124,
-0.04674771428108215,
0.03611365333199501,
-0.0407600961625576,
-0.06515372544527054,
-0.0366581454873085,
0.028253447264432907,
0.036911044269800186,
-0.012107332237064838,
0.13590693473815918,
0.008242727257311344,
0.15984264016151428,
0.14070962369441986,
0.08455470949411392,
-0.062018983066082,
-0.01614982821047306,
0.00374997197650373,
-0.03794719651341438,
0.06668469309806824,
-0.1492520272731781,
0.03878987953066826,
0.12222816795110703,
0.0036256015300750732,
0.15075653791427612,
0.07668574154376984,
-0.0318727046251297,
0.004761490039527416,
0.07289810478687286,
-0.16368791460990906,
-0.10464983433485031,
0.00009764000424183905,
-0.03246529400348663,
-0.11652696877717972,
0.07058043777942657,
0.11345224827528,
-0.07550383359193802,
0.007424657233059406,
-0.008077803067862988,
0.012913945131003857,
-0.043923914432525635,
0.17891693115234375,
0.07228158414363861,
0.04819706082344055,
-0.07201746851205826,
0.07963459193706512,
0.04111414775252342,
-0.0764513835310936,
0.021367210894823074,
0.03768634796142578,
-0.08540183305740356,
-0.049230191856622696,
0.062141843140125275,
0.20156756043434143,
-0.01423653308302164,
-0.06701675802469254,
-0.14322824776172638,
-0.12020532041788101,
0.054041557013988495,
0.1943318098783493,
0.09998999536037445,
0.009111412800848484,
-0.03355679661035538,
0.022769292816519737,
-0.11804656684398651,
0.1169818639755249,
0.0238388329744339,
0.08248627185821533,
-0.15041790902614594,
0.12592314183712006,
0.0026721872854977846,
0.002748612780123949,
-0.026336416602134705,
0.04982507973909378,
-0.1250590682029724,
-0.004575475584715605,
-0.12731711566448212,
-0.00835964921861887,
-0.02422257885336876,
0.006757399998605251,
0.010804131627082825,
-0.048653844743967056,
-0.06720780581235886,
0.01309974305331707,
-0.10238030552864075,
-0.013005602173507214,
0.03758367896080017,
0.0637364462018013,
-0.12436801195144653,
-0.03432577848434448,
0.02284105122089386,
-0.06464679539203644,
0.0618634968996048,
0.013275984674692154,
0.02502279356122017,
0.05850551649928093,
-0.1804674118757248,
0.04051456227898598,
0.07273553311824799,
0.01112209539860487,
0.04299488291144371,
-0.09104887396097183,
-0.008999591693282127,
-0.00036942155566066504,
0.04639844223856926,
0.01837301254272461,
0.09227634966373444,
-0.12612216174602509,
0.00858268141746521,
-0.02911611832678318,
-0.0661216527223587,
-0.04864969849586487,
0.02394390106201172,
0.08970796316862106,
-0.0031197324860841036,
0.2007625550031662,
-0.10177458077669144,
0.007543990854173899,
-0.20263417065143585,
0.008208848536014557,
0.0018247266998514533,
-0.10526265949010849,
-0.12125422060489655,
-0.05762987211346626,
0.04752841591835022,
-0.061031896620988846,
0.15485195815563202,
0.023156948387622833,
0.01696184277534485,
0.035083722323179245,
-0.05255556106567383,
0.040842678397893906,
0.036282457411289215,
0.22782206535339355,
0.024936659261584282,
-0.043650370091199875,
0.017673077061772346,
0.034801073372364044,
0.11213909089565277,
0.07946832478046417,
0.1737278699874878,
0.16970010101795197,
-0.05011652037501335,
0.10503780096769333,
0.05030156672000885,
-0.05557881295681,
-0.12735114991664886,
0.06546703726053238,
-0.05240418389439583,
0.09812047332525253,
-0.027747849002480507,
0.22296196222305298,
0.09065284579992294,
-0.15542227029800415,
0.010374669916927814,
-0.06245582178235054,
-0.08108644187450409,
-0.11855220794677734,
-0.05210752412676811,
-0.09865789115428925,
-0.14928041398525238,
-0.0017190402140840888,
-0.11522171646356583,
-0.0031055316794663668,
0.11036776751279831,
0.0049499934539198875,
-0.015604004263877869,
0.1751014143228531,
0.009885149076581001,
0.04096538946032524,
0.05369431897997856,
0.005636836867779493,
-0.035075843334198,
-0.1082889586687088,
-0.08871903270483017,
0.0020424597896635532,
-0.024779047816991806,
0.020170249044895172,
-0.051712341606616974,
-0.02482708916068077,
0.042481716722249985,
-0.0054224273189902306,
-0.0933804139494896,
0.010526877827942371,
0.02092890441417694,
0.05052768811583519,
0.0449146069586277,
0.009992288425564766,
0.004415431059896946,
0.003955179359763861,
0.20654253661632538,
-0.07617790251970291,
-0.06156919151544571,
-0.10128340870141983,
0.22436833381652832,
0.028622722253203392,
0.026056792587041855,
0.010638391599059105,
-0.08552499115467072,
0.029456347227096558,
0.22967183589935303,
0.18978610634803772,
-0.08123382925987244,
0.0022873501293361187,
-0.0017017981735989451,
-0.011594772338867188,
-0.04123726487159729,
0.10521688312292099,
0.120635025203228,
0.018555501475930214,
-0.07444656640291214,
-0.051881857216358185,
-0.03447159007191658,
-0.0014861669624224305,
-0.04113888368010521,
0.05515667051076889,
0.040599334985017776,
0.015406469814479351,
-0.05036795884370804,
0.05009761452674866,
-0.026512399315834045,
-0.117884561419487,
0.06044561788439751,
-0.19540779292583466,
-0.1470855325460434,
-0.008962666615843773,
0.12081291526556015,
-0.019260087981820107,
0.04905596375465393,
-0.03437152877449989,
-0.0019253732170909643,
0.07329455763101578,
-0.024173526093363762,
-0.06935782730579376,
-0.07961574196815491,
0.05712414160370827,
-0.09850981831550598,
0.23972687125205994,
-0.03788982704281807,
0.04744236543774605,
0.1331075131893158,
0.0435696579515934,
-0.06964361667633057,
0.08705846965312958,
0.04588250443339348,
-0.07699871808290482,
0.024575747549533844,
0.060852356255054474,
-0.04333651810884476,
0.12360280752182007,
0.056070927530527115,
-0.13010545074939728,
0.022732501849532127,
-0.07778707146644592,
-0.09085200726985931,
-0.057102035731077194,
-0.03245355933904648,
-0.06595852971076965,
0.13367566466331482,
0.1871575564146042,
-0.02905338443815708,
0.010387562215328217,
-0.048642612993717194,
0.025183990597724915,
0.06346245110034943,
0.041342705488204956,
-0.03052070178091526,
-0.23528313636779785,
0.0320311076939106,
0.07837603241205215,
-0.010322240181267262,
-0.28996679186820984,
-0.0874861627817154,
-0.00027245268574915826,
-0.04514520615339279,
-0.10071402043104172,
0.07919260859489441,
0.12743087112903595,
0.06120380759239197,
-0.06870241463184357,
-0.10932426899671555,
-0.07340862601995468,
0.15784220397472382,
-0.12739811837673187,
-0.10296729952096939
] |
null | null | ml-agents |
# **ppo** Agent playing **SnowballTarget**
This is a trained model of a **ppo** agent playing **SnowballTarget**
using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
## Usage (with ML-Agents)
The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
- A *longer tutorial* to understand how works ML-Agents:
https://huggingface.co/learn/deep-rl-course/unit5/introduction
### Resume the training
```bash
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
```
### Watch your Agent play
You can watch your agent **playing directly in your browser**
1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
2. Step 1: Find your model_id: r0in/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play 👀
| {"library_name": "ml-agents", "tags": ["SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget"]} | reinforcement-learning | r0in/ppo-SnowballTarget | [
"ml-agents",
"tensorboard",
"onnx",
"SnowballTarget",
"deep-reinforcement-learning",
"reinforcement-learning",
"ML-Agents-SnowballTarget",
"region:us"
] | 2024-02-06T04:59:11+00:00 | [] | [] | TAGS
#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us
|
# ppo Agent playing SnowballTarget
This is a trained model of a ppo agent playing SnowballTarget
using the Unity ML-Agents Library.
## Usage (with ML-Agents)
The Documentation: URL
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your
browser: URL
- A *longer tutorial* to understand how works ML-Agents:
URL
### Resume the training
### Watch your Agent play
You can watch your agent playing directly in your browser
1. If the environment is part of ML-Agents official environments, go to URL
2. Step 1: Find your model_id: r0in/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play
| [
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: r0in/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
"TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n",
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: r0in/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
50,
206
] | [
"passage: TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: r0in/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
-0.03763635456562042,
0.05310651659965515,
-0.0035635484382510185,
0.1111365407705307,
0.17385241389274597,
-0.012339862994849682,
0.13165655732154846,
0.10352768003940582,
0.12891785800457,
0.06043284758925438,
0.09037448465824127,
0.10056623071432114,
0.058443937450647354,
0.12166905403137207,
0.06605803221464157,
-0.22435075044631958,
-0.055104129016399384,
-0.11568691581487656,
-0.005950674880295992,
0.0927736759185791,
0.04745709151029587,
-0.024677056819200516,
0.029173308983445168,
0.05141337960958481,
-0.026634402573108673,
0.0017028588335961103,
-0.06648872792720795,
-0.03470034897327423,
0.0775456503033638,
-0.027190567925572395,
0.01636323519051075,
-0.04024602100253105,
0.11743254959583282,
-0.16557954251766205,
0.029137223958969116,
0.05639517307281494,
-0.016839705407619476,
-0.011153885163366795,
0.14213907718658447,
-0.0006753719644621015,
0.10077476501464844,
-0.10535453259944916,
0.09955383092164993,
0.08563658595085144,
-0.0496867336332798,
-0.004534283187240362,
-0.06801119446754456,
0.06724875420331955,
0.1949647217988968,
0.1491905152797699,
-0.003965415526181459,
0.07199020683765411,
-0.03878626972436905,
0.05972771346569061,
0.17027093470096588,
-0.26547935605049133,
-0.05673579126596451,
0.16734325885772705,
-0.052233077585697174,
0.04350791871547699,
-0.015766650438308716,
0.025137918069958687,
-0.02194664254784584,
0.03302497789263725,
-0.026279347017407417,
0.03774022310972214,
0.27261900901794434,
0.01794952154159546,
-0.09659838676452637,
-0.09260554611682892,
-0.01983162760734558,
0.02925627864897251,
-0.03951714187860489,
-0.16579338908195496,
-0.011930494569242,
0.094725601375103,
0.018077095970511436,
0.04798660799860954,
0.04243338108062744,
0.02023671194911003,
-0.10086513310670853,
-0.1431427150964737,
-0.03329937160015106,
-0.06808651238679886,
0.1016441136598587,
0.10837233066558838,
-0.017126472666859627,
-0.0006278619985096157,
0.04232187941670418,
0.0811476930975914,
0.10821543633937836,
-0.04621521010994911,
-0.03500043973326683,
-0.03439708054065704,
-0.15617237985134125,
0.009801620617508888,
-0.03891485556960106,
0.0033086896874010563,
0.04100678861141205,
0.14701473712921143,
0.15172909200191498,
0.026676753535866737,
0.040478046983480453,
0.03991071134805679,
-0.005376105196774006,
0.1089215949177742,
0.03726876899600029,
-0.007778535597026348,
0.016543693840503693,
0.01330382376909256,
0.049416087567806244,
-0.08809009939432144,
-0.09378399699926376,
0.053175151348114014,
-0.038899216800928116,
0.12510481476783752,
0.15412107110023499,
-0.013204101473093033,
-0.008458150550723076,
-0.032906852662563324,
0.027056189253926277,
-0.15068647265434265,
0.06069779396057129,
0.06873197853565216,
-0.05604860931634903,
-0.08100958913564682,
-0.0476534478366375,
0.03582245483994484,
-0.08025595545768738,
0.03238699957728386,
0.013121218420565128,
0.07128141075372696,
-0.0023862291127443314,
-0.04588692635297775,
0.04095693305134773,
-0.12843330204486847,
-0.007470885757356882,
-0.167567640542984,
-0.11060180515050888,
-0.08844183385372162,
0.031625356525182724,
-0.050912968814373016,
-0.11786554008722305,
-0.10558111220598221,
0.027568098157644272,
-0.0699620246887207,
0.03101528249680996,
-0.034624047577381134,
-0.05864622816443443,
-0.017954671755433083,
-0.11500183492898941,
0.058373093605041504,
0.15964415669441223,
0.015745537355542183,
-0.03342575207352638,
0.020481202751398087,
-0.15151391923427582,
0.17619019746780396,
-0.14873477816581726,
0.13584573566913605,
-0.08593586832284927,
0.035542216151952744,
0.13435718417167664,
-0.030881859362125397,
0.044793255627155304,
0.19158361852169037,
-0.10546379536390305,
-0.07544619590044022,
0.05374712496995926,
-0.0931498259305954,
-0.08616991341114044,
0.06053037568926811,
0.017789874225854874,
0.0316353403031826,
0.0514078289270401,
0.20492717623710632,
0.09157684445381165,
-0.22011323273181915,
0.033746782690286636,
0.011475813575088978,
-0.14302866160869598,
-0.0072789001278579235,
0.13400962948799133,
-0.06597728282213211,
-0.003250439651310444,
-0.03686334937810898,
-0.12963499128818512,
0.10201601684093475,
-0.009882484562695026,
-0.053832296282052994,
0.020083343610167503,
-0.042145173996686935,
-0.037905074656009674,
0.005305909551680088,
0.03134694695472717,
-0.03844805434346199,
-0.06537879258394241,
-0.049959179013967514,
0.04010004922747612,
0.0009177217143587768,
0.07050593197345734,
-0.038737278431653976,
0.11600188910961151,
-0.021803423762321472,
0.014029866084456444,
-0.12742720544338226,
-0.1463773548603058,
-0.02772359736263752,
0.051645174622535706,
0.09389796108007431,
-0.06664573401212692,
0.10614024102687836,
0.0836440771818161,
0.0431043766438961,
-0.08282558619976044,
-0.06654802709817886,
0.012161800637841225,
-0.09668949991464615,
-0.10581791400909424,
-0.07030820101499557,
-0.0664219781756401,
0.1217460036277771,
-0.09078999608755112,
0.06213904172182083,
-0.07566110789775848,
0.08853709697723389,
-0.023493491113185883,
-0.06934677809476852,
0.03501248359680176,
-0.029824066907167435,
0.03606034070253372,
-0.09547371417284012,
0.0994204729795456,
0.07069249451160431,
-0.12458400428295135,
0.040423113852739334,
0.06405740976333618,
-0.08601843565702438,
0.11507050693035126,
0.04290003329515457,
-0.008632710203528404,
-0.04341733455657959,
-0.07373250275850296,
0.013879493810236454,
-0.08621484041213989,
0.024655956774950027,
0.24150678515434265,
0.14444662630558014,
0.08255898952484131,
-0.026803627610206604,
-0.055562254041433334,
-0.030408388003706932,
-0.0628722533583641,
-0.04805012419819832,
0.15865588188171387,
0.04105363041162491,
-0.027097176760435104,
0.0414857491850853,
-0.018954942002892494,
0.08743219822645187,
0.09360624104738235,
-0.00046681068488396704,
-0.1228339672088623,
0.03628643974661827,
0.051405198872089386,
0.05978824198246002,
-0.0011900151148438454,
0.07453135401010513,
-0.018936382606625557,
-0.006726650055497885,
-0.07122646272182465,
-0.012294253334403038,
-0.11710141599178314,
-0.06304767727851868,
0.06468895822763443,
-0.013673163950443268,
-0.023273902013897896,
-0.07331401854753494,
-0.03167157620191574,
0.03604968264698982,
0.08806373178958893,
-0.02498660422861576,
0.0381610281765461,
-0.0403306670486927,
-0.12569159269332886,
0.048725880682468414,
-0.07517753541469574,
-0.22273558378219604,
-0.11282871663570404,
-0.05385291576385498,
-0.06521899253129959,
0.01475948840379715,
0.07546430826187134,
-0.20549245178699493,
-0.0023191282525658607,
-0.08698439598083496,
0.015543486922979355,
-0.019239889457821846,
-0.04120394587516785,
0.12764732539653778,
0.10247287154197693,
-0.01614040695130825,
-0.06139937788248062,
0.010814346373081207,
0.00959622859954834,
-0.07668139040470123,
0.0021899051498621702,
0.06779707968235016,
0.09236805140972137,
0.07436675578355789,
0.04923388361930847,
0.06567050516605377,
-0.018432630226016045,
0.1369025707244873,
-0.041092727333307266,
0.018713917583227158,
0.07168642431497574,
-0.016604673117399216,
0.08294960856437683,
0.021035591140389442,
0.02142866514623165,
0.0067777312360703945,
0.01587519980967045,
0.01369204930961132,
-0.0858071893453598,
-0.23423410952091217,
-0.07913897931575775,
0.008563085459172726,
0.1627383828163147,
0.15867170691490173,
0.10350126773118973,
-0.1296955645084381,
0.02585187554359436,
0.008375695906579494,
-0.12203891575336456,
0.11090033501386642,
0.12336423993110657,
-0.06677190959453583,
-0.01860593818128109,
0.03922714665532112,
-0.04392310604453087,
0.055573076009750366,
0.049977317452430725,
-0.04273345321416855,
0.10752467066049576,
0.0015864752931520343,
-0.006548036355525255,
-0.02070404402911663,
-0.04137907922267914,
-0.03383549675345421,
0.1398027539253235,
0.06528348475694656,
0.03327415511012077,
0.010308478027582169,
-0.047867532819509506,
-0.09526990354061127,
0.13007380068302155,
0.15284639596939087,
-0.07407029718160629,
-0.053534794598817825,
0.1333390474319458,
0.042340535670518875,
0.19962768256664276,
-0.015109330415725708,
-0.1264660358428955,
-0.07000986486673355,
-0.013472472317516804,
-0.11461705714464188,
0.02019449695944786,
0.049600329250097275,
-0.011744379997253418,
-0.1620279997587204,
0.024726856499910355,
-0.010202988050878048,
0.11393574625253677,
0.00015723344404250383,
-0.028692923486232758,
0.06022617593407631,
0.025767778977751732,
-0.022612785920500755,
0.05115843564271927,
-0.1487826108932495,
0.03265663608908653,
-0.013084701262414455,
0.07108616083860397,
-0.061613015830516815,
0.015056871809065342,
0.0897153839468956,
-0.0544009804725647,
0.18246617913246155,
0.04174874722957611,
-0.046953603625297546,
-0.14115479588508606,
-0.17252495884895325,
-0.04399625584483147,
-0.018941396847367287,
-0.11045292019844055,
0.07574015855789185,
0.01927022449672222,
-0.01668671891093254,
-0.09249860048294067,
0.018342195078730583,
-0.028341306373476982,
-0.1348515897989273,
-0.054297368973493576,
-0.09538773447275162,
0.06130342185497284,
-0.05457959324121475,
-0.0720398798584938,
-0.07293558120727539,
0.1821751594543457,
0.09600557386875153,
-0.11171731352806091,
-0.12417850643396378,
0.030175693333148956,
-0.056669652462005615,
-0.04914281517267227,
0.060080233961343765,
0.00651911273598671,
0.10821044445037842,
-0.11013194918632507,
-0.06660923361778259,
-0.02587948553264141,
-0.11143850535154343,
-0.09892358630895615,
0.03670701012015343,
0.16484317183494568,
0.057313140481710434,
0.10225294530391693,
-0.010625135153532028,
0.098576121032238,
-0.01821470633149147,
-0.06735346466302872,
0.13008065521717072,
0.09820667654275894,
-0.028591051697731018,
0.05000705644488335,
0.025326741859316826,
0.0726906955242157,
-0.12224828451871872,
-0.011946289800107479,
0.2409076988697052,
0.26093950867652893,
-0.07177168130874634,
0.2022746503353119,
0.006112528033554554,
-0.03291992470622063,
-0.17216113209724426,
-0.07212060689926147,
0.016152329742908478,
-0.04310839623212814,
0.09142281115055084,
-0.19224762916564941,
0.08063046634197235,
-0.0003387478645890951,
-0.011017072014510632,
0.035620834678411484,
-0.14677974581718445,
-0.06687916815280914,
0.03547535091638565,
0.0985984280705452,
-0.05408632755279541,
-0.0891379788517952,
-0.07231378555297852,
0.014663077890872955,
-0.06962525099515915,
0.0154802855104208,
-0.09540161490440369,
0.06014929711818695,
0.011278577148914337,
0.054993025958538055,
0.056551333516836166,
-0.06285534054040909,
0.10621538758277893,
-0.02926519699394703,
-0.06887596845626831,
-0.06493229418992996,
0.02538309618830681,
-0.032298069447278976,
-0.07722827792167664,
0.0449843592941761,
-0.008350612595677376,
-0.005382488947361708,
-0.1885037124156952,
-0.055947329849004745,
0.006631712894886732,
0.02647165209054947,
-0.02669641189277172,
-0.08123578876256943,
-0.02495962381362915,
0.05929631367325783,
0.09491939842700958,
0.022733641788363457,
0.12070820480585098,
-0.012654382735490799,
-0.0025432382244616747,
0.0725841149687767,
0.022134214639663696,
0.05875968560576439,
-0.13790066540241241,
-0.07043323665857315,
-0.06434459984302521,
-0.006864036433398724,
-0.06678882986307144,
-0.01760229840874672,
0.05605451017618179,
0.06090623512864113,
-0.005553640425205231,
0.06532743573188782,
-0.08528798073530197,
-0.006618343759328127,
0.022178858518600464,
-0.08722984790802002,
-0.12278163433074951,
-0.07596275210380554,
-0.08682159334421158,
0.02227962017059326,
-0.07869023084640503,
0.08333458006381989,
-0.05677209049463272,
-0.012104121036827564,
0.005420186091214418,
0.043917182832956314,
-0.0057458290830254555,
0.04063674807548523,
0.03219658136367798,
0.04190034419298172,
-0.08287359029054642,
0.13101992011070251,
0.007595844101160765,
-0.04083537682890892,
0.04470144584774971,
0.19519077241420746,
-0.06936531513929367,
-0.06653483211994171,
-0.04539110139012337,
0.06274057924747467,
0.017681580036878586,
-0.016959188506007195,
-0.05038421228528023,
-0.05466147139668465,
0.12091242522001266,
-0.16025148332118988,
0.011509301140904427,
-0.11759451776742935,
0.0057615721598267555,
0.04092104360461235,
-0.060975562781095505,
0.06918561458587646,
-0.022071905434131622,
-0.07406055182218552,
-0.1297229379415512,
0.0467652790248394,
0.02632754296064377,
0.10308358073234558,
-0.011803257279098034,
-0.017073096707463264,
-0.1364714354276657,
0.03275436535477638,
-0.010003919713199139,
0.01431594043970108,
-0.14392203092575073,
0.019972005859017372,
-0.011529621668159962,
0.03048848733305931,
0.03981693461537361,
0.06915909051895142,
-0.05229853093624115,
-0.1002628281712532,
-0.05462443456053734,
0.05488888919353485,
-0.0917828306555748,
-0.018075253814458847,
-0.021794764325022697,
-0.07843869179487228,
0.0631992295384407,
0.09321324527263641,
-0.034801363945007324,
-0.03914578631520271,
-0.06366516649723053,
-0.006014901679009199,
-0.02632290869951248,
-0.0495145283639431,
0.059778179973363876,
-0.12880508601665497,
0.03805530071258545,
-0.06751158833503723,
-0.13532404601573944,
0.030565645545721054,
0.11793297529220581,
-0.0673816129565239,
0.06390184164047241,
0.03518886864185333,
-0.07704506814479828,
-0.0878249928355217,
-0.01633692905306816,
0.05653113126754761,
0.07686953991651535,
0.11158570647239685,
-0.08412019908428192,
0.21149738132953644,
-0.1094680204987526,
-0.027788694947957993,
0.024755161255598068,
0.0720515325665474,
0.03544377163052559,
-0.096561960875988,
0.038137901574373245,
-0.01004355400800705,
0.054607003927230835,
0.07496810704469681,
0.01301961112767458,
0.05380402132868767,
0.05718672275543213,
0.12544965744018555,
-0.0023465347476303577,
0.06216884031891823,
-0.0077954987064003944,
0.02609565295279026,
0.1302061527967453,
-0.016735130921006203,
0.06875623762607574,
-0.0788872241973877,
0.08670009672641754,
0.06283020973205566,
0.1061205193400383,
0.05453968793153763,
0.059299685060977936,
-0.09788192808628082,
-0.15804509818553925,
-0.023899666965007782,
0.05553115904331207,
0.015944361686706543,
-0.0478234700858593,
0.194000244140625,
0.14490841329097748,
-0.1811775267124176,
0.015417185612022877,
-0.004899520426988602,
0.04361460730433464,
-0.08378966897726059,
-0.08106022328138351,
-0.001191711169667542,
-0.14553865790367126,
0.11198826134204865,
-0.01352978590875864,
0.0012791699264198542,
-0.00020689047232735902,
0.007799260318279266,
0.03909255564212799,
0.024166187271475792,
-0.03469638526439667,
-0.0004357674333732575,
0.034552477300167084,
-0.03573494032025337,
0.00494375778362155,
0.003148163203150034,
-0.07703456282615662,
-0.022205233573913574,
-0.05859629064798355,
-0.007463590707629919,
0.026189811527729034,
0.0034090119879692793,
0.062386952340602875,
0.001182301202788949,
-0.06382596492767334,
0.06332060694694519,
0.017473623156547546,
0.02264541946351528,
0.2225508689880371,
0.09803426265716553,
-0.04334547370672226,
-0.04769124463200569,
0.2195543646812439,
-0.03931450471282005,
-0.06647414714097977,
-0.08758091181516647,
0.12667861580848694,
-0.04447319731116295,
-0.04399633780121803,
-0.031007224693894386,
-0.17426174879074097,
-0.058614592999219894,
0.16199305653572083,
0.13106590509414673,
-0.024161070585250854,
0.006118630059063435,
-0.06359899789094925,
0.005020173732191324,
0.020252373069524765,
0.08908343315124512,
0.08208692073822021,
0.060135163366794586,
-0.1221841499209404,
-0.015538173727691174,
-0.08418500423431396,
-0.09640301764011383,
-0.17353950440883636,
0.05290071293711662,
0.0219383854418993,
-0.020914731547236443,
-0.020668432116508484,
0.12145189195871353,
-0.12254182994365692,
-0.11314327269792557,
0.09930120408535004,
-0.031330306082963943,
-0.07344083487987518,
-0.004925206303596497,
0.024922242388129234,
0.016233354806900024,
0.12297514826059341,
0.08325483649969101,
0.04406316205859184,
0.0369301363825798,
-0.018536413088440895,
-0.10657919943332672,
0.041539791971445084,
0.04230421781539917,
-0.11673042923212051,
0.22539131343364716,
-0.01880943961441517,
-0.0013480670750141144,
0.09627034515142441,
0.06006057187914848,
-0.18408727645874023,
0.020528141409158707,
0.043877940624952316,
-0.18613897264003754,
0.011535394005477428,
0.1053571105003357,
-0.04032641276717186,
0.009284698404371738,
0.052258048206567764,
-0.04840974882245064,
0.018356211483478546,
0.170683816075325,
0.03514442592859268,
-0.04043722152709961,
0.07821901142597198,
-0.13788215816020966,
0.09926414489746094,
0.10703399777412415,
-0.0532216802239418,
0.0157996267080307,
-0.05513165518641472,
0.0014267932856455445,
0.0026730920653790236,
-0.019001923501491547,
-0.014829971827566624,
-0.1037287712097168,
-0.023600658401846886,
-0.04076157882809639,
0.025302108377218246,
-0.22316965460777283,
-0.12580741941928864,
-0.053945623338222504,
-0.08078240603208542,
-0.0512552373111248,
0.08562391996383667,
0.07313624024391174,
-0.035351529717445374,
0.01786971651017666,
-0.10460549592971802,
0.031099172309041023,
0.14883935451507568,
-0.0867714211344719,
-0.000036686888051917776
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-classification | dylansheils0241/Quantum-Balanced-GPT2-Experimental-Theoretical-Classifier-Arxiv-V2 | [
"transformers",
"safetensors",
"gpt2",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T05:01:01+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #gpt2 #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #gpt2 #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
57,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #gpt2 #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.056688208132982254,
0.16275818645954132,
-0.004995849449187517,
0.02452681213617325,
0.09786820411682129,
0.012234015390276909,
0.0665494054555893,
0.11446622014045715,
-0.019654832780361176,
0.11005201935768127,
0.03224760666489601,
0.09871180355548859,
0.11353186517953873,
0.16168983280658722,
0.0032966197468340397,
-0.23542851209640503,
0.0468909777700901,
-0.12772662937641144,
-0.03536919504404068,
0.1168198511004448,
0.1453043669462204,
-0.10465475916862488,
0.07777879387140274,
-0.030737513676285744,
-0.00966490525752306,
-0.031576599925756454,
-0.06401672959327698,
-0.05254754796624184,
0.05026170611381531,
0.07030009478330612,
0.06729904562234879,
0.007110965438187122,
0.09183509647846222,
-0.26528099179267883,
0.02310207113623619,
0.07594640552997589,
-0.001530517591163516,
0.07526049017906189,
0.04812141880393028,
-0.08188801258802414,
0.07724089175462723,
-0.05910906940698624,
0.14970065653324127,
0.08168394863605499,
-0.08909645676612854,
-0.19279919564723969,
-0.08510304987430573,
0.10391399264335632,
0.18374687433242798,
0.05092503875494003,
-0.023054618388414383,
0.10542254149913788,
-0.08633118867874146,
0.009634695015847683,
0.05877980589866638,
-0.06533382833003998,
-0.05344819277524948,
0.07030105590820312,
0.07782267779111862,
0.07990992069244385,
-0.1312698870897293,
-0.015684951096773148,
0.011104214005172253,
0.00718331104144454,
0.08438988029956818,
0.021947048604488373,
0.13984841108322144,
0.04317888617515564,
-0.13002023100852966,
-0.043909087777137756,
0.09818204492330551,
0.04373921453952789,
-0.05486512929201126,
-0.2536156177520752,
-0.024169430136680603,
-0.04035896062850952,
-0.030552461743354797,
-0.03810810670256615,
0.046126991510391235,
-0.011335964314639568,
0.08135324716567993,
-0.01014806516468525,
-0.08107352256774902,
-0.039619747549295425,
0.06577171385288239,
0.06208868697285652,
0.026821019127964973,
-0.011671994812786579,
0.010234266519546509,
0.12182370573282242,
0.11196604371070862,
-0.1267545372247696,
-0.057587433606386185,
-0.06730430573225021,
-0.08423306047916412,
-0.0447598397731781,
0.033019017428159714,
0.043012093752622604,
0.043217264115810394,
0.24644720554351807,
0.009206846356391907,
0.056639526039361954,
0.046275459229946136,
0.01094451081007719,
0.06766996532678604,
0.11699596792459488,
-0.06068361550569534,
-0.09169819205999374,
-0.028210563585162163,
0.09167566895484924,
0.00641299644485116,
-0.04253393039107323,
-0.06145499274134636,
0.064153291285038,
0.012870057485997677,
0.12154073268175125,
0.08118180185556412,
0.003567382926121354,
-0.07309924066066742,
-0.06970501691102982,
0.1891379952430725,
-0.16031186282634735,
0.048485323786735535,
0.031144030392169952,
-0.04012530297040939,
-0.0037560793571174145,
0.010843836702406406,
0.023869646713137627,
-0.02002854458987713,
0.09128747135400772,
-0.05557823181152344,
-0.03632063418626785,
-0.11016727983951569,
-0.032307349145412445,
0.031602974981069565,
0.005229081027209759,
-0.027400631457567215,
-0.034758999943733215,
-0.088483065366745,
-0.05970429629087448,
0.09185574948787689,
-0.0725538581609726,
-0.05133562162518501,
-0.014709736220538616,
-0.07693814486265182,
0.023501750081777573,
0.017334258183836937,
0.08247776329517365,
-0.02171487919986248,
0.03736530989408493,
-0.05768737196922302,
0.0629277229309082,
0.11147940158843994,
0.028959784656763077,
-0.053054340183734894,
0.05642477050423622,
-0.24564996361732483,
0.10391399264335632,
-0.07363908737897873,
0.050764378160238266,
-0.1495596021413803,
-0.026108751073479652,
0.04620397463440895,
0.008229508064687252,
-0.013179546222090721,
0.14200609922409058,
-0.2165803462266922,
-0.03158329799771309,
0.1675405651330948,
-0.09414954483509064,
-0.07587666064500809,
0.058981720358133316,
-0.05404539033770561,
0.10750386118888855,
0.044090528041124344,
-0.02138153277337551,
0.06216038018465042,
-0.14397647976875305,
-0.012432611547410488,
-0.041649285703897476,
-0.02405562996864319,
0.16330546140670776,
0.07457496225833893,
-0.06195339933037758,
0.08059893548488617,
0.02354867197573185,
-0.01621856912970543,
-0.04427330568432808,
-0.022748108953237534,
-0.10862526297569275,
0.010284756310284138,
-0.060920849442481995,
0.025102287530899048,
-0.024612700566649437,
-0.09277556091547012,
-0.029063016176223755,
-0.17864122986793518,
-0.006479884497821331,
0.0886693224310875,
-0.011500544846057892,
-0.021484078839421272,
-0.12116942554712296,
0.009813597425818443,
0.03316587954759598,
0.0017007269198074937,
-0.13188430666923523,
-0.05852383002638817,
0.027688656002283096,
-0.16301771998405457,
0.03372528776526451,
-0.05495554581284523,
0.05014612898230553,
0.03457905724644661,
-0.03310814127326012,
-0.0332946851849556,
0.009236995130777359,
0.006859376560896635,
-0.012469877488911152,
-0.2507966458797455,
-0.023342570289969444,
-0.022434204816818237,
0.1754746288061142,
-0.22036518156528473,
0.04243079200387001,
0.07755454629659653,
0.1488277018070221,
0.00758343143388629,
-0.0385059118270874,
0.0108621371909976,
-0.081815205514431,
-0.030218740925192833,
-0.06505976617336273,
-0.012460521422326565,
-0.03667744621634483,
-0.05610157549381256,
0.0502706803381443,
-0.15897555649280548,
-0.02874920517206192,
0.1054336428642273,
0.06867773085832596,
-0.14192140102386475,
-0.019301993772387505,
-0.04151778295636177,
-0.043261781334877014,
-0.059273406863212585,
-0.052341289818286896,
0.12165731191635132,
0.05511542409658432,
0.04830443114042282,
-0.06941264122724533,
-0.07721688598394394,
0.007358690723776817,
-0.01781853660941124,
-0.01777339167892933,
0.0893767699599266,
0.07364588975906372,
-0.1244131475687027,
0.09259122610092163,
0.10102657228708267,
0.09181178361177444,
0.1081208810210228,
-0.018081961199641228,
-0.08706964552402496,
-0.044923022389411926,
0.0203016996383667,
0.013586522080004215,
0.14873908460140228,
-0.013833221048116684,
0.056480638682842255,
0.037942372262477875,
-0.011916344054043293,
0.011820046231150627,
-0.09332224726676941,
0.02717048116028309,
0.03382128104567528,
-0.01951325498521328,
0.0386216901242733,
-0.041371311992406845,
0.02094746008515358,
0.08781760185956955,
0.04612930491566658,
0.043133337050676346,
0.0117363091558218,
-0.04232318326830864,
-0.10922972857952118,
0.17441187798976898,
-0.1265207678079605,
-0.24811580777168274,
-0.14101167023181915,
0.003548931796103716,
0.04161277040839195,
-0.00981870200484991,
0.0033757472410798073,
-0.07208114862442017,
-0.11677295714616776,
-0.09369951486587524,
0.014811713248491287,
0.04634920507669449,
-0.07345830649137497,
-0.059883084148168564,
0.06119617074728012,
0.03956949710845947,
-0.1438819319009781,
0.022251952439546585,
0.04618961736559868,
-0.08916490525007248,
-0.007660253439098597,
0.08383868634700775,
0.0663260892033577,
0.17821210622787476,
0.006946447771042585,
-0.0211239792406559,
0.030622977763414383,
0.21115799248218536,
-0.14198173582553864,
0.11189346015453339,
0.13871541619300842,
-0.0914137065410614,
0.08057228475809097,
0.19499939680099487,
0.039358362555503845,
-0.10546562820672989,
0.03305496647953987,
0.025585372000932693,
-0.028574291616678238,
-0.25104397535324097,
-0.0688476711511612,
0.0032778505701571703,
-0.04926806315779686,
0.07188434898853302,
0.09186343103647232,
0.0980982705950737,
0.014891031198203564,
-0.0960264801979065,
-0.08064629137516022,
0.047425657510757446,
0.1060817539691925,
0.0058129834942519665,
-0.014884848147630692,
0.08918209373950958,
-0.033973969519138336,
0.01830443926155567,
0.09493532031774521,
0.0027011006604880095,
0.1717296540737152,
0.05488133803009987,
0.17306490242481232,
0.07624487578868866,
0.06581548601388931,
0.01999000459909439,
0.005975092761218548,
0.02183385193347931,
0.017030013725161552,
-0.0042428807355463505,
-0.08410344272851944,
-0.006547621451318264,
0.11983280628919601,
0.06439385563135147,
0.025222649797797203,
0.01179569773375988,
-0.03775259107351303,
0.08442632108926773,
0.1749287247657776,
0.0019723710138350725,
-0.18502026796340942,
-0.07310080528259277,
0.07859374582767487,
-0.09147641062736511,
-0.10472535341978073,
-0.033703628927469254,
0.03427765890955925,
-0.17517809569835663,
0.022020921111106873,
-0.017561664804816246,
0.1075635626912117,
-0.14025665819644928,
-0.018815070390701294,
0.06322184205055237,
0.07119301706552505,
-0.003150238422676921,
0.06206432729959488,
-0.15356871485710144,
0.10289750248193741,
0.01623719558119774,
0.06722601503133774,
-0.09720844030380249,
0.10133114457130432,
-0.004952651914209127,
-0.004525967873632908,
0.13827085494995117,
0.01132762711495161,
-0.05856676027178764,
-0.09017363935709,
-0.10513435304164886,
-0.00838832464069128,
0.12849153578281403,
-0.15005145967006683,
0.08388647437095642,
-0.02834545075893402,
-0.04350460320711136,
0.0028875069692730904,
-0.12221316248178482,
-0.13001364469528198,
-0.18832695484161377,
0.056929487735033035,
-0.13712386786937714,
0.03853238746523857,
-0.10552185028791428,
-0.04352074861526489,
-0.02887987159192562,
0.20707611739635468,
-0.23298580944538116,
-0.06692921370267868,
-0.15351387858390808,
-0.08341002464294434,
0.14223980903625488,
-0.047327857464551926,
0.08947230875492096,
-0.0006640679202973843,
0.18996402621269226,
0.023749830201268196,
-0.022958843037486076,
0.1032608300447464,
-0.0953274741768837,
-0.19397945702075958,
-0.08692745119333267,
0.1546938568353653,
0.1391226202249527,
0.03686843439936638,
-0.004141106735914946,
0.02847231552004814,
-0.016847684979438782,
-0.124631866812706,
0.024518294259905815,
0.18073658645153046,
0.08076948672533035,
0.0265391543507576,
-0.027251770719885826,
-0.11268990486860275,
-0.06502450257539749,
-0.03222431242465973,
0.02961815521121025,
0.18243654072284698,
-0.07519686967134476,
0.1901955008506775,
0.13443516194820404,
-0.05532145872712135,
-0.19614705443382263,
0.020869113504886627,
0.040682703256607056,
0.010022376663982868,
0.03179483115673065,
-0.20703792572021484,
0.09156925976276398,
0.0010500495554879308,
-0.05144895613193512,
0.13911594450473785,
-0.17432743310928345,
-0.151635080575943,
0.06508468836545944,
0.03710223361849785,
-0.1928347796201706,
-0.11609705537557602,
-0.08698900043964386,
-0.04764305055141449,
-0.17105866968631744,
0.10657104849815369,
0.01810937561094761,
0.00958965066820383,
0.03534632921218872,
0.024905970320105553,
0.011848334223031998,
-0.038653843104839325,
0.1937946081161499,
-0.025732165202498436,
0.030279111117124557,
-0.08093483000993729,
-0.06396064907312393,
0.05981128662824631,
-0.05653243884444237,
0.07926622778177261,
-0.019695239141583443,
0.013051961548626423,
-0.11097230017185211,
-0.04620685055851936,
-0.0259378794580698,
0.01774062030017376,
-0.09204115718603134,
-0.09366398304700851,
-0.052516400814056396,
0.09481028467416763,
0.08910299837589264,
-0.03606134280562401,
-0.037287481129169464,
-0.07352357357740402,
0.033029962331056595,
0.1711474508047104,
0.17759248614311218,
0.03955881670117378,
-0.08047898858785629,
-0.0056870789267122746,
-0.012103362940251827,
0.04403502494096756,
-0.22654037177562714,
0.06325183808803558,
0.04530617222189903,
0.01543486025184393,
0.11385229974985123,
-0.01837165653705597,
-0.1614435911178589,
-0.06635874509811401,
0.06131202355027199,
-0.06802738457918167,
-0.18228155374526978,
0.003949065692722797,
0.059882380068302155,
-0.16518694162368774,
-0.03804792836308479,
0.04252162575721741,
-0.005406341515481472,
-0.04330739378929138,
0.016907572746276855,
0.08069808781147003,
0.005998523905873299,
0.07098331302404404,
0.05882091447710991,
0.08421110361814499,
-0.1037181094288826,
0.07749924808740616,
0.08153212815523148,
-0.08422624319791794,
0.03190286085009575,
0.08996295183897018,
-0.06251151114702225,
-0.031116170808672905,
0.03407644107937813,
0.07890257239341736,
0.021274501457810402,
-0.042060818523168564,
0.007018920499831438,
-0.09714305400848389,
0.06359373033046722,
0.097147636115551,
0.03615867346525192,
0.014581901021301746,
0.033455245196819305,
0.04942217841744423,
-0.07525192946195602,
0.11710844188928604,
0.022087544202804565,
0.018151897937059402,
-0.04620935395359993,
-0.053078390657901764,
0.03228947892785072,
-0.022450316697359085,
-0.00941105093806982,
-0.03882573917508125,
-0.07127608358860016,
-0.018803894519805908,
-0.1570902317762375,
-0.012990092858672142,
-0.04015745967626572,
0.008910941891372204,
0.025404607877135277,
-0.036834996193647385,
0.00830900203436613,
0.01183675043284893,
-0.070340096950531,
-0.06075016036629677,
-0.022623594850301743,
0.09379690140485764,
-0.16431091725826263,
0.02377411536872387,
0.08262428641319275,
-0.12145231664180756,
0.09788227826356888,
0.02060251496732235,
0.0030464455485343933,
0.027605198323726654,
-0.15334390103816986,
0.04065665975213051,
-0.02279796451330185,
0.012528573162853718,
0.04368005692958832,
-0.22050756216049194,
0.0022407027427107096,
-0.03955530747771263,
-0.0581013448536396,
-0.009252764284610748,
-0.02985531836748123,
-0.11230431497097015,
0.09915203601121902,
0.008330761454999447,
-0.08242792636156082,
-0.031244277954101562,
0.03579036891460419,
0.08468461036682129,
-0.025901922956109047,
0.15200486779212952,
0.0009069927618838847,
0.07067206501960754,
-0.17469091713428497,
-0.02179952897131443,
-0.011881532147526741,
0.022119969129562378,
-0.014708246104419231,
-0.015323184430599213,
0.0425114743411541,
-0.024193845689296722,
0.1896541714668274,
-0.024643851444125175,
0.03960045427083969,
0.0647975355386734,
0.01552598923444748,
-0.014887680299580097,
0.11118700355291367,
0.05733522027730942,
0.022819433361291885,
0.021128227934241295,
0.009885239414870739,
-0.03985453024506569,
-0.013325306586921215,
-0.19498084485530853,
0.06593974679708481,
0.13870294392108917,
0.08520747721195221,
-0.011946371756494045,
0.07486340403556824,
-0.10040875524282455,
-0.12106827646493912,
0.11215335875749588,
-0.06267691403627396,
-0.008283826522529125,
-0.06529007107019424,
0.1326436847448349,
0.14720159769058228,
-0.19087740778923035,
0.06812075525522232,
-0.06286486238241196,
-0.054577119648456573,
-0.11731726676225662,
-0.19296373426914215,
-0.056126803159713745,
-0.0537908673286438,
-0.015372869558632374,
-0.04924733191728592,
0.07244181632995605,
0.056062083691358566,
0.011923623271286488,
0.0036824082490056753,
0.07538694143295288,
-0.01878892257809639,
0.001183385495096445,
0.031768519431352615,
0.06635171920061111,
0.012988004833459854,
-0.030890334397554398,
0.019896110519766808,
-0.008031128905713558,
0.040116652846336365,
0.06276603788137436,
0.03778861090540886,
-0.02647724375128746,
0.016218312084674835,
-0.03651279956102371,
-0.10915597528219223,
0.042008984833955765,
-0.029179582372307777,
-0.08002755790948868,
0.13762354850769043,
0.0252593494951725,
0.005759078077971935,
-0.021546276286244392,
0.24652670323848724,
-0.07283888757228851,
-0.08912359923124313,
-0.14807496964931488,
0.10262012481689453,
-0.03890256956219673,
0.06634368002414703,
0.040788695216178894,
-0.10723847150802612,
0.014206798747181892,
0.12468050420284271,
0.15960600972175598,
-0.04598904773592949,
0.020886680111289024,
0.03171207755804062,
0.0034098350442945957,
-0.03958895057439804,
0.053595054894685745,
0.06899604946374893,
0.1463599056005478,
-0.048738062381744385,
0.08973133563995361,
0.005670548416674137,
-0.10229548066854477,
-0.03761244937777519,
0.11828185617923737,
-0.01828082464635372,
0.019166678190231323,
-0.05744493380188942,
0.11868125200271606,
-0.0608985461294651,
-0.23095959424972534,
0.06514488160610199,
-0.07566448301076889,
-0.14436253905296326,
-0.021428024396300316,
0.07276028394699097,
-0.01780269481241703,
0.026690328493714333,
0.07398071140050888,
-0.07777206599712372,
0.1971588432788849,
0.03926048055291176,
-0.05780704692006111,
-0.05835490673780441,
0.07752813398838043,
-0.1160687804222107,
0.28119364380836487,
0.011715840548276901,
0.05096624419093132,
0.1037345826625824,
-0.015980001538991928,
-0.1366477608680725,
0.01282057911157608,
0.09203975647687912,
-0.08909518271684647,
0.038567181676626205,
0.21223965287208557,
-0.003803485771641135,
0.11778628826141357,
0.07744387537240982,
-0.07308902591466904,
0.04619969055056572,
-0.11599848419427872,
-0.07299992442131042,
-0.08707643300294876,
0.09514311701059341,
-0.07195685058832169,
0.14144541323184967,
0.12355200946331024,
-0.052837129682302475,
0.011311785317957401,
-0.0302458256483078,
0.045323196798563004,
0.00835395511239767,
0.10012959688901901,
0.010363605804741383,
-0.18085741996765137,
0.024026650935411453,
0.015367760322988033,
0.10938689112663269,
-0.17194917798042297,
-0.09750598669052124,
0.0472152978181839,
0.001750203431583941,
-0.0601421594619751,
0.12909571826457977,
0.056909650564193726,
0.04916413500905037,
-0.0436386875808239,
-0.017397185787558556,
-0.009748851880431175,
0.13655178248882294,
-0.1076984852552414,
-0.0029465733096003532
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# huner_ncbi_disease_dslim
This model is a fine-tuned version of [dslim/distilbert-NER](https://huggingface.co/dslim/distilbert-NER) on the transformer_dataset_ner dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1484
- Precision: 0.8325
- Recall: 0.8653
- F1: 0.8486
- Accuracy: 0.9850
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.1243 | 1.0 | 667 | 0.0669 | 0.7013 | 0.8412 | 0.7649 | 0.9787 |
| 0.0512 | 2.0 | 1334 | 0.0656 | 0.7825 | 0.8412 | 0.8108 | 0.9818 |
| 0.0221 | 3.0 | 2001 | 0.0744 | 0.7908 | 0.8501 | 0.8194 | 0.9822 |
| 0.0107 | 4.0 | 2668 | 0.1022 | 0.7940 | 0.8475 | 0.8199 | 0.9808 |
| 0.008 | 5.0 | 3335 | 0.1055 | 0.7818 | 0.8602 | 0.8191 | 0.9816 |
| 0.0057 | 6.0 | 4002 | 0.1173 | 0.8067 | 0.8590 | 0.832 | 0.9830 |
| 0.0027 | 7.0 | 4669 | 0.1188 | 0.8188 | 0.8501 | 0.8342 | 0.9834 |
| 0.0022 | 8.0 | 5336 | 0.1229 | 0.8080 | 0.8450 | 0.8261 | 0.9826 |
| 0.0019 | 9.0 | 6003 | 0.1341 | 0.8007 | 0.8526 | 0.8258 | 0.9834 |
| 0.0019 | 10.0 | 6670 | 0.1360 | 0.8045 | 0.8628 | 0.8326 | 0.9822 |
| 0.0011 | 11.0 | 7337 | 0.1376 | 0.8163 | 0.8640 | 0.8395 | 0.9838 |
| 0.0008 | 12.0 | 8004 | 0.1447 | 0.8007 | 0.8577 | 0.8282 | 0.9833 |
| 0.0006 | 13.0 | 8671 | 0.1381 | 0.8139 | 0.8615 | 0.8370 | 0.9839 |
| 0.0005 | 14.0 | 9338 | 0.1398 | 0.8297 | 0.8666 | 0.8477 | 0.9843 |
| 0.0004 | 15.0 | 10005 | 0.1404 | 0.8232 | 0.8640 | 0.8431 | 0.9842 |
| 0.0003 | 16.0 | 10672 | 0.1486 | 0.8329 | 0.8551 | 0.8439 | 0.9838 |
| 0.0 | 17.0 | 11339 | 0.1469 | 0.8114 | 0.8691 | 0.8393 | 0.9837 |
| 0.0002 | 18.0 | 12006 | 0.1500 | 0.8297 | 0.8602 | 0.8447 | 0.9843 |
| 0.0001 | 19.0 | 12673 | 0.1489 | 0.8315 | 0.8653 | 0.8481 | 0.9849 |
| 0.0 | 20.0 | 13340 | 0.1484 | 0.8325 | 0.8653 | 0.8486 | 0.9850 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["transformer_dataset_ner"], "metrics": ["precision", "recall", "f1", "accuracy"], "base_model": "dslim/distilbert-NER", "model-index": [{"name": "huner_ncbi_disease_dslim", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "transformer_dataset_ner", "type": "transformer_dataset_ner", "config": "ncbi_disease", "split": "validation", "args": "ncbi_disease"}, "metrics": [{"type": "precision", "value": 0.8325183374083129, "name": "Precision"}, {"type": "recall", "value": 0.8653113087674714, "name": "Recall"}, {"type": "f1", "value": 0.8485981308411215, "name": "F1"}, {"type": "accuracy", "value": 0.9849891909996041, "name": "Accuracy"}]}]}]} | token-classification | manibt1993/huner_ncbi_disease_dslim | [
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"token-classification",
"generated_from_trainer",
"dataset:transformer_dataset_ner",
"base_model:dslim/distilbert-NER",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T05:05:05+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #distilbert #token-classification #generated_from_trainer #dataset-transformer_dataset_ner #base_model-dslim/distilbert-NER #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| huner\_ncbi\_disease\_dslim
===========================
This model is a fine-tuned version of dslim/distilbert-NER on the transformer\_dataset\_ner dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1484
* Precision: 0.8325
* Recall: 0.8653
* F1: 0.8486
* Accuracy: 0.9850
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 20
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #distilbert #token-classification #generated_from_trainer #dataset-transformer_dataset_ner #base_model-dslim/distilbert-NER #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
87,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #token-classification #generated_from_trainer #dataset-transformer_dataset_ner #base_model-dslim/distilbert-NER #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.11331743746995926,
0.16518492996692657,
-0.0032087506260722876,
0.10786605626344681,
0.10086948424577713,
-0.010844642296433449,
0.15423667430877686,
0.1319924294948578,
-0.06677702069282532,
0.09634270519018173,
0.15692844986915588,
0.1441306322813034,
0.014604879543185234,
0.184182807803154,
-0.061467502266168594,
-0.2052789032459259,
0.043272633105516434,
0.045440852642059326,
-0.04191911593079567,
0.11982046067714691,
0.09290284663438797,
-0.1324814110994339,
0.09846729785203934,
0.02540888451039791,
-0.18621741235256195,
-0.012294299900531769,
0.019675718620419502,
-0.06333060562610626,
0.09771048277616501,
0.008025059476494789,
0.10377772897481918,
0.04232848808169365,
0.05380699038505554,
-0.18050427734851837,
0.007092174608260393,
0.05585828423500061,
0.004727312363684177,
0.10541149228811264,
0.030744915828108788,
-0.02346157841384411,
0.02765505202114582,
-0.10174203664064407,
0.06372696906328201,
0.02911604382097721,
-0.12050726264715195,
-0.25632187724113464,
-0.10513312369585037,
0.09693165123462677,
0.0732484832406044,
0.0573834553360939,
-0.00524371862411499,
0.1544465720653534,
-0.015539906919002533,
0.09228827059268951,
0.2447313368320465,
-0.30619069933891296,
-0.06164926290512085,
0.04534580186009407,
0.024167625233530998,
0.07387848198413849,
-0.10456188023090363,
-0.014678644016385078,
0.06584122776985168,
0.022174660116434097,
0.16938161849975586,
-0.009241725318133831,
0.03696867823600769,
-0.015325741842389107,
-0.1401745229959488,
-0.05743548646569252,
0.17606444656848907,
0.06717025488615036,
-0.051546610891819,
-0.08488064259290695,
-0.07689113169908524,
-0.17347589135169983,
-0.03583303466439247,
-0.0430050827562809,
0.044941578060388565,
-0.035892002284526825,
-0.06627263128757477,
-0.013122932985424995,
-0.07906093448400497,
-0.05923003330826759,
-0.022124335169792175,
0.14565499126911163,
0.05136236548423767,
-0.0014297518646344543,
-0.012046732939779758,
0.08102662861347198,
-0.018798764795064926,
-0.15525129437446594,
-0.012624857947230339,
0.0062922267243266106,
0.014009237289428711,
-0.04267289116978645,
-0.03163391351699829,
-0.08191222697496414,
0.026309704408049583,
0.1811678111553192,
-0.05866526439785957,
0.05600323900580406,
0.008720449171960354,
0.0369781069457531,
-0.0913560539484024,
0.16439957916736603,
-0.045586489140987396,
-0.007782112341374159,
0.027987709268927574,
0.09800636768341064,
0.06459930539131165,
-0.010780341923236847,
-0.10182953625917435,
0.041016045957803726,
0.13984067738056183,
0.011234750971198082,
-0.03557255491614342,
0.07372425496578217,
-0.06041949987411499,
-0.027430560439825058,
0.06647209078073502,
-0.10116682201623917,
0.01785271428525448,
-0.0068526491522789,
-0.04583390802145004,
-0.04873289540410042,
0.012510150671005249,
0.002598557621240616,
0.00948102306574583,
0.057379063218832016,
-0.10855495184659958,
-0.014163322746753693,
-0.055451828986406326,
-0.10585574805736542,
0.02167429029941559,
-0.11159665137529373,
0.007080341223627329,
-0.10610157251358032,
-0.14798593521118164,
-0.006046136375516653,
0.05283717066049576,
-0.0240136981010437,
-0.05816395953297615,
-0.05607003718614578,
-0.08490588515996933,
0.029642639681696892,
-0.0012113285483792424,
0.008063274435698986,
-0.05108615383505821,
0.07620904594659805,
0.05254064127802849,
0.07764924317598343,
-0.015048994682729244,
0.03295058757066727,
-0.09523802995681763,
0.07323917001485825,
-0.20562493801116943,
0.03884311765432358,
-0.07776899635791779,
0.07764405012130737,
-0.11841346323490143,
-0.08313498646020889,
-0.0004979028599336743,
-0.03303254395723343,
0.07045215368270874,
0.10523517429828644,
-0.1600155085325241,
-0.07061348110437393,
0.18803730607032776,
-0.09699844568967819,
-0.14529408514499664,
0.12028022110462189,
-0.04637552425265312,
0.03353828936815262,
0.05734061077237129,
0.22181861102581024,
0.08518590033054352,
-0.07785896956920624,
-0.043912384659051895,
-0.03514569625258446,
0.053206831216812134,
-0.04662558436393738,
0.07246474921703339,
-0.004500572104007006,
0.037550024688243866,
0.02493489347398281,
-0.0561881959438324,
0.030427780002355576,
-0.06920983642339706,
-0.08856954425573349,
-0.05829475447535515,
-0.09307615458965302,
0.07276712357997894,
0.0400504507124424,
0.0648331418633461,
-0.10519187152385712,
-0.0942566767334938,
0.06176988407969475,
0.08338073641061783,
-0.08183879405260086,
0.03186149150133133,
-0.09512895345687866,
0.11693070828914642,
-0.1029515191912651,
-0.014793370850384235,
-0.1537422090768814,
-0.03320751711726189,
0.04320092871785164,
-0.05704433470964432,
-0.004327652510255575,
-0.02886890061199665,
0.07374884933233261,
0.06715530902147293,
-0.05601140111684799,
-0.05735449120402336,
-0.0336463525891304,
0.02417094074189663,
-0.10182850807905197,
-0.19662319123744965,
-0.03140781447291374,
-0.041312627494335175,
0.11371716111898422,
-0.20465591549873352,
0.0569358728826046,
0.053539715707302094,
0.13042303919792175,
0.05304742231965065,
-0.03974863886833191,
-0.003516195574775338,
0.036346886307001114,
-0.04771659895777702,
-0.08254354447126389,
0.04452196881175041,
0.014833367429673672,
-0.10144075006246567,
-0.022815635427832603,
-0.13362173736095428,
0.1636219173669815,
0.11216863244771957,
0.01649344153702259,
-0.06993233412504196,
0.002426037099212408,
-0.05357388034462929,
-0.03852064535021782,
-0.014812429435551167,
0.008136647753417492,
0.09543093293905258,
0.006247631274163723,
0.1441926509141922,
-0.09581916034221649,
-0.05239937826991081,
0.04871254414319992,
-0.04564985632896423,
-0.0249228123575449,
0.09873908758163452,
0.023037342354655266,
-0.12876927852630615,
0.16368292272090912,
0.163777157664299,
-0.06642480194568634,
0.11098010092973709,
-0.06395170092582703,
-0.05262422934174538,
-0.05045084282755852,
0.02489553764462471,
0.03660368174314499,
0.11416584253311157,
-0.06706155836582184,
-0.004517789930105209,
0.02380867674946785,
0.019939212128520012,
-0.005358528345823288,
-0.17232076823711395,
-0.0045275031588971615,
0.04403099790215492,
-0.05461524799466133,
0.0194020364433527,
-0.00246609328314662,
-0.02321595512330532,
0.07781235873699188,
0.012133797630667686,
-0.05870119109749794,
0.056238166987895966,
0.013416665606200695,
-0.06667135655879974,
0.19259107112884521,
-0.08861493319272995,
-0.15016771852970123,
-0.1301594078540802,
-0.04176505282521248,
-0.08203121274709702,
0.01858595944941044,
0.056667156517505646,
-0.07575559616088867,
-0.03674240782856941,
-0.10569609701633453,
-0.04204271361231804,
0.02491126023232937,
0.03502808138728142,
0.05153784155845642,
-0.019807426258921623,
0.10896241664886475,
-0.10407214611768723,
-0.017356297001242638,
-0.024242984130978584,
0.004383387975394726,
0.03782741352915764,
0.00938501488417387,
0.11481577903032303,
0.11918585002422333,
-0.013177754357457161,
0.027436228469014168,
-0.016598090529441833,
0.22586242854595184,
-0.05407344922423363,
-0.004810855258256197,
0.11952225118875504,
-0.007186289876699448,
0.08024009317159653,
0.13519428670406342,
0.04326054826378822,
-0.09046954661607742,
0.0010775300906971097,
0.027095258235931396,
-0.03402423486113548,
-0.20156870782375336,
-0.024219879880547523,
-0.04010973125696182,
0.0040936884470283985,
0.11880958080291748,
0.04180790111422539,
0.07357224076986313,
0.06560362130403519,
0.011900858022272587,
0.062316056340932846,
-0.01646030880510807,
0.10428105294704437,
0.11472908407449722,
0.05098157376050949,
0.1228976622223854,
-0.04201100021600723,
-0.04129574075341225,
0.02390459179878235,
0.02002076432108879,
0.22140805423259735,
0.011765853501856327,
0.21555742621421814,
0.05777095630764961,
0.18092601001262665,
0.015149841085076332,
0.05345592275261879,
-0.010599667206406593,
-0.025677785277366638,
-0.0013693073997274041,
-0.05111465975642204,
-0.030413802713155746,
0.036900609731674194,
-0.061276670545339584,
0.07359965890645981,
-0.08460884541273117,
0.049172692000865936,
0.06705562025308609,
0.2694116532802582,
0.045531149953603745,
-0.36065834760665894,
-0.10501615703105927,
0.005136039573699236,
-0.016589978709816933,
-0.030821729451417923,
0.006974308285862207,
0.11773128062486649,
-0.05663339048624039,
0.056596655398607254,
-0.08560369908809662,
0.06915554404258728,
-0.05073726177215576,
0.029830338433384895,
0.03588331490755081,
0.09468097984790802,
-0.003548147389665246,
0.053040217608213425,
-0.26000910997390747,
0.2472548633813858,
0.02273217961192131,
0.06407202780246735,
-0.04067898541688919,
0.008073920384049416,
0.014844098128378391,
0.07871391624212265,
0.10300952941179276,
0.0000591035932302475,
-0.08949214220046997,
-0.1670103818178177,
-0.10813657194375992,
0.00024229644623119384,
0.08040017634630203,
-0.04802706837654114,
0.10204843431711197,
-0.019305948168039322,
-0.004541144706308842,
0.056905657052993774,
-0.0030684303492307663,
-0.06906034052371979,
-0.10135903209447861,
0.00980785209685564,
0.03936152532696724,
-0.029731934890151024,
-0.09952665120363235,
-0.10034690797328949,
-0.08753369003534317,
0.17677733302116394,
-0.06147194281220436,
-0.051653750240802765,
-0.12417303770780563,
0.04787381365895271,
0.06557279080152512,
-0.09988472610712051,
0.053104255348443985,
-0.015145414508879185,
0.13567253947257996,
0.007477185223251581,
-0.058403294533491135,
0.10647305101156235,
-0.05794433504343033,
-0.18432168662548065,
-0.04712490364909172,
0.1198895052075386,
0.024958541616797447,
0.037753377109766006,
-0.001469952636398375,
0.03415736183524132,
-0.002714614849537611,
-0.06937306374311447,
0.04013996571302414,
0.04067183658480644,
0.06397997587919235,
-0.03740749880671501,
-0.01972205750644207,
0.021592767909169197,
-0.06534962356090546,
-0.008431735448539257,
0.15332354605197906,
0.2665109634399414,
-0.10217682272195816,
0.03596844896674156,
0.04409985989332199,
-0.05106870457530022,
-0.17153136432170868,
0.023274345323443413,
0.051005151122808456,
-0.0018602911150082946,
0.029337124899029732,
-0.14232195913791656,
0.08413314819335938,
0.09483026713132858,
-0.02766880951821804,
0.07083029299974442,
-0.26802533864974976,
-0.1295289695262909,
0.10596908628940582,
0.13912136852741241,
0.08976710587739944,
-0.13412334024906158,
-0.06527061015367508,
-0.013783574104309082,
-0.1153033897280693,
0.10086079686880112,
-0.08614299446344376,
0.09495245665311813,
-0.02846066653728485,
0.031416382640600204,
0.008488121442496777,
-0.060370881110429764,
0.12649841606616974,
0.01981380768120289,
0.09523286670446396,
-0.047916658222675323,
-0.01781081221997738,
0.10198492556810379,
-0.0925862193107605,
0.04582023620605469,
-0.1129106655716896,
0.0544205866754055,
-0.06486762315034866,
-0.025303559377789497,
-0.06331204622983932,
0.033821951597929,
-0.03546653315424919,
-0.05107218027114868,
-0.056474532932043076,
0.026838883757591248,
0.07578695565462112,
-0.008395163342356682,
0.195644810795784,
0.041985414922237396,
0.13294512033462524,
0.17115581035614014,
0.06078561395406723,
-0.06833864748477936,
-0.06837835162878036,
-0.016837673261761665,
-0.036426883190870285,
0.055092062801122665,
-0.17303793132305145,
0.050497762858867645,
0.11853437125682831,
0.01688811555504799,
0.13101860880851746,
0.050240129232406616,
-0.05669824033975601,
-0.0003212453448213637,
0.05712943524122238,
-0.16327019035816193,
-0.14998435974121094,
0.0005219808663241565,
-0.01613185741007328,
-0.14945180714130402,
0.0804029181599617,
0.11294721066951752,
-0.0759836807847023,
-0.006838797591626644,
0.0006191277061589062,
0.014619335532188416,
-0.015506677329540253,
0.16862988471984863,
0.07461019605398178,
0.0667327344417572,
-0.08441136032342911,
0.08269046992063522,
0.050886351615190506,
-0.06019144505262375,
0.03381671383976936,
0.0077400729060173035,
-0.11136568337678909,
-0.020914670079946518,
0.03163399547338486,
0.15929605066776276,
-0.0165549386292696,
-0.04530994966626167,
-0.1641664057970047,
-0.09854574501514435,
0.05710165947675705,
0.14995217323303223,
0.09332187473773956,
0.0360635407269001,
-0.027492955327033997,
-0.0028005526401102543,
-0.11360589414834976,
0.12793882191181183,
0.04751303791999817,
0.10515379905700684,
-0.18585999310016632,
0.08204816281795502,
-0.009040947072207928,
0.020293163135647774,
-0.01452384889125824,
0.03213182091712952,
-0.10314235836267471,
-0.018811846151947975,
-0.09612733125686646,
0.010510391555726528,
-0.058329883962869644,
0.000784470874350518,
-0.007742663379758596,
-0.05382144823670387,
-0.0559295229613781,
0.03514036163687706,
-0.09222658723592758,
-0.03696984425187111,
0.037106651812791824,
0.05559393763542175,
-0.12192552536725998,
-0.02322584204375744,
0.02433006279170513,
-0.08616817742586136,
0.07865220308303833,
0.02013666369020939,
0.016287868842482567,
0.023197075352072716,
-0.07125385850667953,
0.0358576774597168,
0.06616940349340439,
0.004403811879456043,
0.04512833431363106,
-0.12533949315547943,
-0.024008948355913162,
0.000013652723282575607,
0.001961507136002183,
0.012021160684525967,
0.07564131170511246,
-0.126017227768898,
-0.004657834768295288,
-0.029813453555107117,
-0.030544761568307877,
-0.06948251277208328,
0.03957745060324669,
0.10160690546035767,
0.007052895147353411,
0.2037791907787323,
-0.07629746198654175,
0.006295789033174515,
-0.19844451546669006,
-0.00135095976293087,
0.0073311785236001015,
-0.12886576354503632,
-0.09152509272098541,
-0.022987017408013344,
0.03695394843816757,
-0.06483880430459976,
0.11927477270364761,
-0.03673669695854187,
-0.03539114072918892,
0.03958427160978317,
-0.043730031698942184,
0.00620079692453146,
0.027116650715470314,
0.19685497879981995,
0.012861435301601887,
-0.0443960465490818,
0.06419001519680023,
0.009519466198980808,
0.09315811097621918,
0.08507039397954941,
0.14260029792785645,
0.1762513667345047,
-0.03424835950136185,
0.08762777596712112,
0.03783570975065231,
-0.019475502893328667,
-0.1762257069349289,
0.10205364227294922,
-0.061336666345596313,
0.10378770530223846,
0.014517568983137608,
0.17340923845767975,
0.1351848989725113,
-0.1637544482946396,
0.021666523069143295,
-0.016964567825198174,
-0.08615066111087799,
-0.0918741226196289,
-0.1286894679069519,
-0.10240145027637482,
-0.15392646193504333,
-0.0006990717956796288,
-0.10751817375421524,
-0.0012221930082887411,
0.06239936500787735,
-0.0005214822012931108,
-0.012537149712443352,
0.17764411866664886,
0.031041910871863365,
0.03235291317105293,
0.04429911822080612,
-0.005892420653253794,
-0.07842524349689484,
-0.05172451213002205,
-0.0837441235780716,
0.018247978761792183,
0.003156965598464012,
0.04281877726316452,
-0.033967096358537674,
0.010846452787518501,
0.042700499296188354,
-0.01454508863389492,
-0.10928541421890259,
0.012012404389679432,
0.018684910610318184,
0.023947719484567642,
0.00966023001819849,
0.019259603694081306,
-0.016176333650946617,
-0.012571067549288273,
0.16275186836719513,
-0.05998019501566887,
-0.025029076263308525,
-0.11740328371524811,
0.18437282741069794,
0.037282418459653854,
-0.011116410605609417,
0.024632718414068222,
-0.08275814354419708,
0.024418560788035393,
0.166922464966774,
0.15421055257320404,
-0.00776335084810853,
0.0054455227218568325,
-0.023107541725039482,
-0.01798749342560768,
-0.03712504357099533,
0.05826127901673317,
0.08958557993173599,
-0.022374119609594345,
-0.04973643645644188,
-0.032802898436784744,
-0.05166168883442879,
-0.02390911802649498,
-0.051199816167354584,
0.06346315145492554,
0.021285157650709152,
0.007934019900858402,
-0.05945749208331108,
0.03494865074753761,
-0.03245242312550545,
-0.07360716164112091,
0.06440584361553192,
-0.18307094275951385,
-0.14755654335021973,
-0.018828263506293297,
0.07464199513196945,
-0.01668204553425312,
0.0389910526573658,
-0.008807429112493992,
-0.0016689473995938897,
0.0814555361866951,
-0.015442125499248505,
-0.06823352724313736,
-0.061284806579351425,
0.06573106348514557,
-0.10559387505054474,
0.22098283469676971,
-0.030079450458288193,
0.02130650170147419,
0.13345767557621002,
0.02701803855597973,
-0.12068349123001099,
0.06271307170391083,
0.04636808857321739,
-0.03078523464500904,
0.008786688558757305,
0.1023842915892601,
-0.01674838736653328,
0.10901854932308197,
0.05363907292485237,
-0.11387653648853302,
-0.02249734476208687,
-0.05228453502058983,
-0.04341258853673935,
-0.06137155741453171,
-0.03286110982298851,
-0.04669284448027611,
0.13104474544525146,
0.16488583385944366,
-0.058179520070552826,
-0.010202362202107906,
-0.043402306735515594,
0.032388556748628616,
0.08793600648641586,
0.01230678427964449,
-0.01778518781065941,
-0.2556462585926056,
0.030764618888497353,
0.05488218739628792,
-0.0036565980408340693,
-0.24672706425189972,
-0.11423309892416,
-0.008063092827796936,
-0.048756688833236694,
-0.06936703622341156,
0.08792514353990555,
0.10354989767074585,
0.0619996003806591,
-0.063043013215065,
-0.023283228278160095,
-0.08698628842830658,
0.15572118759155273,
-0.13144811987876892,
-0.07963345944881439
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# Aanshula/layoutlm-funsd-tf
This model is a fine-tuned version of [microsoft/layoutlm-base-uncased](https://huggingface.co/microsoft/layoutlm-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.3182
- Validation Loss: 0.6807
- Train Overall Precision: 0.7172
- Train Overall Recall: 0.7878
- Train Overall F1: 0.7508
- Train Overall Accuracy: 0.7864
- Epoch: 6
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 3e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: mixed_float16
### Training results
| Train Loss | Validation Loss | Train Overall Precision | Train Overall Recall | Train Overall F1 | Train Overall Accuracy | Epoch |
|:----------:|:---------------:|:-----------------------:|:--------------------:|:----------------:|:----------------------:|:-----:|
| 1.7000 | 1.4167 | 0.2445 | 0.2107 | 0.2264 | 0.4831 | 0 |
| 1.1656 | 0.8677 | 0.5749 | 0.6257 | 0.5992 | 0.7251 | 1 |
| 0.7704 | 0.7254 | 0.6356 | 0.7160 | 0.6734 | 0.7637 | 2 |
| 0.5758 | 0.6690 | 0.6851 | 0.7476 | 0.7150 | 0.7857 | 3 |
| 0.4526 | 0.6096 | 0.7085 | 0.7757 | 0.7406 | 0.8046 | 4 |
| 0.3614 | 0.6834 | 0.7118 | 0.7657 | 0.7377 | 0.7872 | 5 |
| 0.3182 | 0.6807 | 0.7172 | 0.7878 | 0.7508 | 0.7864 | 6 |
### Framework versions
- Transformers 4.35.2
- TensorFlow 2.15.0
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_keras_callback"], "base_model": "microsoft/layoutlm-base-uncased", "model-index": [{"name": "Aanshula/layoutlm-funsd-tf", "results": []}]} | token-classification | Aanshula/layoutlm-funsd-tf | [
"transformers",
"tf",
"tensorboard",
"layoutlm",
"token-classification",
"generated_from_keras_callback",
"base_model:microsoft/layoutlm-base-uncased",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T05:06:14+00:00 | [] | [] | TAGS
#transformers #tf #tensorboard #layoutlm #token-classification #generated_from_keras_callback #base_model-microsoft/layoutlm-base-uncased #license-mit #autotrain_compatible #endpoints_compatible #region-us
| Aanshula/layoutlm-funsd-tf
==========================
This model is a fine-tuned version of microsoft/layoutlm-base-uncased on an unknown dataset.
It achieves the following results on the evaluation set:
* Train Loss: 0.3182
* Validation Loss: 0.6807
* Train Overall Precision: 0.7172
* Train Overall Recall: 0.7878
* Train Overall F1: 0.7508
* Train Overall Accuracy: 0.7864
* Epoch: 6
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* optimizer: {'name': 'AdamWeightDecay', 'learning\_rate': 3e-05, 'decay': 0.0, 'beta\_1': 0.9, 'beta\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight\_decay\_rate': 0.01}
* training\_precision: mixed\_float16
### Training results
### Framework versions
* Transformers 4.35.2
* TensorFlow 2.15.0
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': 3e-05, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: mixed\\_float16",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tf #tensorboard #layoutlm #token-classification #generated_from_keras_callback #base_model-microsoft/layoutlm-base-uncased #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': 3e-05, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: mixed\\_float16",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
74,
122,
4,
31
] | [
"passage: TAGS\n#transformers #tf #tensorboard #layoutlm #token-classification #generated_from_keras_callback #base_model-microsoft/layoutlm-base-uncased #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': 3e-05, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: mixed\\_float16### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09931270778179169,
0.09231573343276978,
-0.0037152450531721115,
0.07736225426197052,
0.08962064236402512,
0.036955755203962326,
0.15003326535224915,
0.15483999252319336,
-0.09336654841899872,
0.07638853043317795,
0.12876538932323456,
0.13035272061824799,
0.0568365678191185,
0.22059322893619537,
-0.10841826349496841,
-0.17790383100509644,
0.044364020228385925,
0.0182371623814106,
-0.052003588527441025,
0.09191381931304932,
0.08985824882984161,
-0.10284438729286194,
0.10938577353954315,
0.004476784262806177,
-0.20042073726654053,
0.014705892652273178,
0.07311666756868362,
-0.08470035344362259,
0.08020906895399094,
0.07495155930519104,
0.08026311546564102,
0.040614087134599686,
0.03616613149642944,
-0.13853713870048523,
0.012771411798894405,
0.08320941776037216,
-0.014211083762347698,
0.1181720495223999,
0.03384261205792427,
-0.013161573559045792,
0.060493264347314835,
-0.11802928149700165,
0.03688130900263786,
0.04590633139014244,
-0.12706060707569122,
-0.24402961134910583,
-0.1302802711725235,
0.06478685885667801,
0.05240360647439957,
0.02632199600338936,
0.007795820012688637,
0.21155835688114166,
0.007995794527232647,
0.10166133940219879,
0.22616013884544373,
-0.3687729835510254,
-0.051643721759319305,
0.0798247829079628,
0.023527463898062706,
0.000646707892883569,
-0.07327625155448914,
0.04235582798719406,
0.053464632481336594,
0.02814042940735817,
0.10831150412559509,
-0.03268716111779213,
-0.020840879529714584,
-0.025735842064023018,
-0.10602171719074249,
-0.01485949382185936,
0.1769445389509201,
0.05450713634490967,
-0.06809885799884796,
-0.07769507169723511,
-0.06371089816093445,
-0.15520229935646057,
-0.021420907229185104,
-0.0528961680829525,
0.03640866279602051,
-0.024449186399579048,
-0.1022203341126442,
-0.06771481037139893,
-0.06721439957618713,
-0.0519077368080616,
-0.0929550752043724,
0.13833008706569672,
0.011331267654895782,
0.061889681965112686,
-0.028559427708387375,
0.08283065259456635,
-0.07826177030801773,
-0.12931974232196808,
-0.029709642753005028,
-0.017043540254235268,
-0.062411829829216,
-0.05739511176943779,
-0.05232281982898712,
-0.11205431818962097,
0.053206317126750946,
0.13999725878238678,
-0.06998135894536972,
0.0749567374587059,
-0.08046231418848038,
0.031559836119413376,
-0.05577186867594719,
0.11534234136343002,
-0.040406860411167145,
-0.010358343832194805,
0.028368273749947548,
0.03452272713184357,
0.056944381445646286,
-0.02540329471230507,
-0.08639465272426605,
-0.002021809108555317,
0.09677024185657501,
0.018654881045222282,
-0.021286070346832275,
0.09152480214834213,
-0.06906471401453018,
-0.045915260910987854,
0.03915739804506302,
-0.10721520334482193,
0.0315583311021328,
-0.023288272321224213,
-0.08779802918434143,
-0.0015806615119799972,
0.063624806702137,
0.010142842307686806,
-0.02567620575428009,
0.029268790036439896,
-0.08657465130090714,
-0.017490820959210396,
-0.0734746977686882,
-0.1243511363863945,
0.027037840336561203,
-0.14180605113506317,
-0.02151241898536682,
-0.07556012272834778,
-0.19698801636695862,
-0.0014614496612921357,
0.061471693217754364,
-0.06496231257915497,
0.0034349486231803894,
-0.04182056710124016,
-0.1098404973745346,
0.045601051300764084,
-0.019512586295604706,
0.09068566560745239,
-0.050167180597782135,
0.05753815546631813,
0.018176237121224403,
0.08365822583436966,
-0.10859450697898865,
0.025705935433506966,
-0.04816235974431038,
0.03805983066558838,
-0.2281416654586792,
0.09164304286241531,
-0.06964509189128876,
0.029571568593382835,
-0.1481555551290512,
-0.05999341607093811,
0.01593293435871601,
-0.018956897780299187,
0.0948294922709465,
0.10880745947360992,
-0.21028892695903778,
-0.050617340952157974,
0.14927230775356293,
-0.10664462298154831,
-0.11984732002019882,
0.08764691650867462,
-0.044129565358161926,
0.034804049879312515,
0.07778709381818771,
0.16975219547748566,
-0.003532606875523925,
-0.12071144580841064,
-0.022778958082199097,
-0.020717639476060867,
-0.014330863952636719,
-0.01979137398302555,
0.03809812664985657,
-0.022268738597631454,
0.016068993136286736,
0.012739313766360283,
0.0011769456323236227,
0.010204690508544445,
-0.07997783273458481,
-0.059295494109392166,
-0.06003418192267418,
-0.06476195901632309,
0.08012580126523972,
0.01980697177350521,
0.06223805993795395,
-0.09602485597133636,
-0.09841927886009216,
0.11714772135019302,
0.054395467042922974,
-0.044686079025268555,
0.03274887427687645,
-0.11661546677350998,
0.06339741498231888,
-0.05449041724205017,
-0.006059584207832813,
-0.1863192468881607,
-0.09736741334199905,
0.016071295365691185,
0.025070467963814735,
0.022562537342309952,
-0.010644628666341305,
0.08263859152793884,
0.02945917285978794,
-0.07654811441898346,
0.007755253929644823,
0.005374782718718052,
0.017523229122161865,
-0.09676339477300644,
-0.24954071640968323,
-0.05055242031812668,
-0.04034411534667015,
0.053320273756980896,
-0.200605109333992,
0.02784215658903122,
0.08789078146219254,
0.17565619945526123,
0.09398511797189713,
-0.03412085399031639,
0.013290227390825748,
0.02513348124921322,
-0.03354423865675926,
-0.08284997195005417,
0.03448433056473732,
0.028898421674966812,
-0.14948460459709167,
-0.002649195957928896,
-0.13922119140625,
0.15615810453891754,
0.14082539081573486,
-0.009285302832722664,
-0.10746833682060242,
-0.02080707438290119,
-0.04135835915803909,
-0.028090111911296844,
-0.01811150275170803,
0.023531725630164146,
0.07812994718551636,
0.021343057975172997,
0.13024111092090607,
-0.07191052287817001,
-0.05994518846273422,
0.042674720287323,
-0.03499416261911392,
-0.039722491055727005,
0.05995939299464226,
-0.04280836135149002,
-0.19432906806468964,
0.11876805871725082,
0.12027589231729507,
-0.041069090366363525,
0.1575523465871811,
-0.05323069170117378,
-0.06291639804840088,
-0.07850142568349838,
0.012570283375680447,
0.046168915927410126,
0.11667497456073761,
-0.061277102679014206,
-0.00265197129920125,
0.03894735500216484,
0.016915250569581985,
-0.009417137131094933,
-0.15276598930358887,
0.012708980590105057,
-0.01106188166886568,
-0.07118801027536392,
0.01725820079445839,
0.029627760872244835,
0.01284532155841589,
0.12009972333908081,
0.029442008584737778,
0.004695577081292868,
0.07635868340730667,
0.0012593315914273262,
-0.08091410249471664,
0.18608437478542328,
-0.11643136292695999,
-0.1302950084209442,
-0.10225316137075424,
-0.063622385263443,
-0.10149458795785904,
-0.015897681936621666,
0.012505761347711086,
-0.0765141025185585,
-0.054232094436883926,
-0.08273349702358246,
-0.034669842571020126,
0.01001502014696598,
0.061675623059272766,
0.049999501556158066,
0.003907068632543087,
0.11703499406576157,
-0.09903082251548767,
-0.02982174977660179,
-0.04247487336397171,
-0.013891729526221752,
0.022847842425107956,
0.014348599128425121,
0.048077285289764404,
0.09015308320522308,
-0.03871561214327812,
0.06075994297862053,
-0.06786133348941803,
0.2044920176267624,
-0.04255348816514015,
0.017228923738002777,
0.11794155836105347,
-0.035921648144721985,
0.051100391894578934,
0.11537431180477142,
0.0520748645067215,
-0.11046956479549408,
0.014613429084420204,
0.05309632793068886,
-0.03663588687777519,
-0.24701087176799774,
-0.025340616703033447,
-0.04291250929236412,
-0.0417473129928112,
0.06032927706837654,
0.038633525371551514,
0.11301629990339279,
0.038688283413648605,
0.02043928951025009,
0.09437335282564163,
-0.010479573160409927,
0.08537095785140991,
0.1423676609992981,
0.06032809242606163,
0.12061893194913864,
-0.0688808411359787,
0.024223797023296356,
0.07898861169815063,
0.021514222025871277,
0.22757573425769806,
0.010619520209729671,
0.12527957558631897,
0.07094694674015045,
0.11589638888835907,
0.009921222925186157,
0.028260452672839165,
0.007340989075601101,
-0.024025753140449524,
0.005528825335204601,
-0.07033969461917877,
-0.00975480955094099,
0.04538164660334587,
-0.03138028457760811,
0.06376045942306519,
-0.07955615222454071,
0.057207126170396805,
0.07307484745979309,
0.26801520586013794,
0.07729794830083847,
-0.35497772693634033,
-0.10044755041599274,
0.01226211991161108,
0.011735524982213974,
-0.04219532385468483,
-0.03994675725698471,
0.09974009543657303,
-0.06006893143057823,
0.15368573367595673,
-0.08847745507955551,
0.06536833196878433,
-0.04658222571015358,
0.016743583604693413,
0.06903325021266937,
0.12132571637630463,
-0.01567535474896431,
0.008187981322407722,
-0.29843220114707947,
0.29040247201919556,
0.06892320513725281,
0.10174385458230972,
-0.06327513605356216,
0.013564834371209145,
0.029612766578793526,
0.04129265621304512,
0.11323575675487518,
-0.013618797063827515,
-0.14691297709941864,
-0.11981874704360962,
-0.08842993527650833,
-0.0025085932575166225,
0.10728497803211212,
0.03666640818119049,
0.08813938498497009,
-0.016126243397593498,
0.00026025224360637367,
0.06223839521408081,
-0.045250698924064636,
-0.16297809779644012,
-0.05593502148985863,
0.03111671842634678,
0.06742115318775177,
-0.06678187847137451,
-0.06794218719005585,
-0.09906315803527832,
-0.0437074713408947,
0.24670752882957458,
-0.0699869766831398,
-0.062130093574523926,
-0.15322870016098022,
0.13420474529266357,
0.09135541319847107,
-0.058200377970933914,
0.04717684164643288,
-0.011315431445837021,
0.10216126590967178,
0.04191255941987038,
-0.14000624418258667,
0.153783917427063,
-0.035267360508441925,
-0.1803128868341446,
-0.05436774343252182,
0.09581082314252853,
0.03745904564857483,
0.04816357418894768,
0.005944895092397928,
0.05033399164676666,
0.05945215001702309,
-0.07792054861783981,
0.11013174802064896,
0.044924501329660416,
0.05223962664604187,
0.01350854616612196,
-0.00513118552044034,
-0.08380395919084549,
-0.016818877309560776,
0.004052373580634594,
0.14999200403690338,
0.24231284856796265,
-0.11334749311208725,
0.06608209758996964,
0.011244718916714191,
-0.07821674644947052,
-0.17757315933704376,
0.10113973915576935,
0.0828341543674469,
0.007002583704888821,
-0.007575619034469128,
-0.15493705868721008,
0.04814807325601578,
0.0971434935927391,
-0.020711373537778854,
0.09530018270015717,
-0.25599220395088196,
-0.15002411603927612,
0.08217590302228928,
0.10159792751073837,
0.11385046690702438,
-0.14342249929904938,
-0.07711153477430344,
-0.031553708016872406,
-0.08786807954311371,
0.14138765633106232,
-0.15892337262630463,
0.10023172199726105,
-0.004500150680541992,
0.02424776926636696,
0.005755405407398939,
-0.041938282549381256,
0.10551048815250397,
0.0003228755376767367,
0.10721466690301895,
-0.06191457435488701,
-0.014592454768717289,
0.13387344777584076,
-0.09395262598991394,
0.07199355214834213,
-0.024647127836942673,
0.04627736285328865,
-0.0791526585817337,
0.0020325391087681055,
-0.05896372348070145,
0.06837870925664902,
-0.024025432765483856,
-0.03899528458714485,
-0.035159144550561905,
0.04867102578282356,
0.03182580694556236,
-0.04128335788846016,
0.16917186975479126,
0.0029910379089415073,
0.17768594622612,
0.2242049127817154,
0.10185476392507553,
-0.05885549634695053,
-0.006097961682826281,
0.04383426904678345,
-0.054635174572467804,
0.05507711321115494,
-0.1801450401544571,
0.04046925529837608,
0.11813580989837646,
0.006359335035085678,
0.12127089500427246,
0.056042760610580444,
-0.07008528709411621,
0.05293973535299301,
0.07049693912267685,
-0.17139697074890137,
-0.10378953069448471,
0.020267445594072342,
-0.045507144182920456,
-0.11339105665683746,
0.10872401297092438,
0.158878892660141,
-0.03069385513663292,
0.01955552212893963,
0.014919859357178211,
0.019115161150693893,
-0.05583059415221214,
0.1479482352733612,
0.027124326676130295,
0.04579480364918709,
-0.10149062424898148,
0.13203631341457367,
0.028881127014756203,
-0.05558861047029495,
0.09279289096593857,
0.03342122212052345,
-0.09943683445453644,
-0.009230552241206169,
0.006078067701309919,
0.15284046530723572,
-0.03346576169133186,
-0.05205067992210388,
-0.1277771294116974,
-0.12212744355201721,
0.07267900556325912,
0.25025367736816406,
0.05601974576711655,
0.04852888733148575,
-0.026896126568317413,
-0.015220362693071365,
-0.06340610980987549,
0.06560417264699936,
0.05808901786804199,
0.07120687514543533,
-0.16627831757068634,
0.13678112626075745,
-0.00527710048481822,
0.027097944170236588,
-0.04058877006173134,
0.02035738341510296,
-0.14325003325939178,
-0.015191109851002693,
-0.12167292088270187,
0.004551813006401062,
-0.0470723919570446,
-0.011881469748914242,
0.005568870808929205,
-0.07251667976379395,
-0.07469169795513153,
0.03868599981069565,
-0.09365430474281311,
-0.027056992053985596,
0.050557758659124374,
0.029998984187841415,
-0.141539067029953,
-0.04888519272208214,
-0.024321094155311584,
-0.06936667859554291,
0.07604549825191498,
0.03981931135058403,
0.0015536609571427107,
0.015469584614038467,
-0.0644650086760521,
-0.005351127125322819,
0.07172442972660065,
-0.016471605747938156,
0.08578913658857346,
-0.10116153210401535,
-0.01437048614025116,
-0.00585296330973506,
0.048105526715517044,
0.036827072501182556,
0.14567448198795319,
-0.09586511552333832,
-0.03161715716123581,
-0.038071077316999435,
-0.001011099899187684,
-0.06437652558088303,
0.07670499384403229,
0.16254635155200958,
0.016645362600684166,
0.15240991115570068,
-0.09535875916481018,
-0.020931683480739594,
-0.20191170275211334,
0.007290266919881105,
0.001627088407985866,
-0.11694607138633728,
-0.07478068023920059,
0.0031893705017864704,
0.08893660455942154,
-0.08453892916440964,
0.13747544586658478,
-0.02282123826444149,
0.031870804727077484,
0.061517808586359024,
-0.0651489719748497,
-0.08572006225585938,
0.039822548627853394,
0.18184535205364227,
0.027485361322760582,
-0.022374041378498077,
0.061415430158376694,
0.023810049518942833,
0.0640149638056755,
0.0494888611137867,
0.2241557389497757,
0.0865892842411995,
-0.025702808052301407,
0.12874768674373627,
0.07715126127004623,
-0.039682332426309586,
-0.13682879507541656,
0.1409902274608612,
-0.07238047569990158,
0.1563117653131485,
-0.023603567853569984,
0.05517289415001869,
0.10391518473625183,
-0.17435196042060852,
0.019633954390883446,
-0.025698687881231308,
-0.0824364498257637,
-0.12244835495948792,
-0.11841907352209091,
-0.10172849893569946,
-0.1403558850288391,
-0.006115767173469067,
-0.1217241957783699,
0.04430493712425232,
0.04884900897741318,
0.03438449278473854,
0.0206530150026083,
0.11401882767677307,
-0.042009469121694565,
0.02317008376121521,
0.09674979001283646,
0.005149711389094591,
-0.046943746507167816,
-0.06989170610904694,
-0.06064256280660629,
0.05073654279112816,
0.036176860332489014,
0.026420576497912407,
0.003892711363732815,
0.026691190898418427,
0.05641726404428482,
-0.01748221181333065,
-0.11700664460659027,
0.02083597332239151,
0.055385325103998184,
0.02385678142309189,
0.01903678849339485,
0.0411616750061512,
-0.020685236901044846,
-0.02977241761982441,
0.1451166868209839,
-0.09546728432178497,
-0.018091140314936638,
-0.17193537950515747,
0.217141255736351,
-0.007771022152155638,
-0.0061240289360284805,
0.020138323307037354,
-0.08927961438894272,
-0.007445264607667923,
0.14978493750095367,
0.19460096955299377,
-0.016191942617297173,
-0.01125924102962017,
0.016019757837057114,
-0.012763028033077717,
-0.034012943506240845,
0.11716088652610779,
0.05117005482316017,
-0.025992101058363914,
-0.05165327340364456,
-0.058856938034296036,
-0.02935018576681614,
-0.03107297047972679,
-0.017486829310655594,
0.05714653804898262,
0.028177743777632713,
0.0037913552951067686,
-0.025378374382853508,
0.07066759467124939,
-0.04614965245127678,
-0.10260120034217834,
0.07632248848676682,
-0.19940434396266937,
-0.1597406566143036,
-0.01108819805085659,
0.008496991358697414,
-0.003148365532979369,
0.04394904524087906,
-0.024989329278469086,
0.02094740979373455,
0.10713931918144226,
-0.03198446333408356,
-0.07873263210058212,
-0.09768203645944595,
0.07705888152122498,
-0.09306227415800095,
0.16065707802772522,
-0.02674105577170849,
0.04174595698714256,
0.12906184792518616,
0.017859414219856262,
-0.12203430384397507,
0.015428747981786728,
0.025962013751268387,
-0.04684153571724892,
0.006177021190524101,
0.09625044465065002,
-0.020453685894608498,
0.13353043794631958,
0.052997685968875885,
-0.09956391155719757,
-0.0029926735442131758,
-0.0782589465379715,
-0.07052016258239746,
-0.05431295558810234,
-0.017110653221607208,
-0.060124680399894714,
0.1331445425748825,
0.1960698962211609,
-0.041408099234104156,
0.010629183612763882,
-0.05052262544631958,
0.029262658208608627,
0.08163203299045563,
-0.03084740974009037,
-0.04280102625489235,
-0.2149379551410675,
0.06716424971818924,
0.09967762976884842,
0.010472824797034264,
-0.20044848322868347,
-0.08262143284082413,
-0.0019835978746414185,
-0.02701849862933159,
-0.08610531687736511,
0.09288732707500458,
0.10394030064344406,
0.068637415766716,
-0.05723019689321518,
-0.08486308157444,
-0.013187723234295845,
0.12616902589797974,
-0.12543830275535583,
-0.06472274661064148
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# safety-utcustom-train-SF-RGBD-b5
This model is a fine-tuned version of [nvidia/mit-b5](https://huggingface.co/nvidia/mit-b5) on the sam1120/safety-utcustom-TRAIN dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0867
- Mean Iou: 0.7280
- Mean Accuracy: 0.7762
- Overall Accuracy: 0.9818
- Accuracy Unlabeled: nan
- Accuracy Safe: 0.5578
- Accuracy Unsafe: 0.9947
- Iou Unlabeled: nan
- Iou Safe: 0.4745
- Iou Unsafe: 0.9814
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 4e-06
- train_batch_size: 15
- eval_batch_size: 15
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 120
### Training results
| Training Loss | Epoch | Step | Accuracy Safe | Accuracy Unlabeled | Accuracy Unsafe | Iou Safe | Iou Unlabeled | Iou Unsafe | Validation Loss | Mean Accuracy | Mean Iou | Overall Accuracy |
|:-------------:|:------:|:----:|:-------------:|:------------------:|:---------------:|:--------:|:-------------:|:----------:|:---------------:|:-------------:|:--------:|:----------------:|
| 0.789 | 0.91 | 10 | 0.0203 | nan | 0.8957 | 0.0095 | 0.0 | 0.8722 | 0.9555 | 0.4580 | 0.2939 | 0.8698 |
| 0.7579 | 1.82 | 20 | 0.0117 | nan | 0.9614 | 0.0069 | 0.0 | 0.9338 | 0.8322 | 0.4866 | 0.3136 | 0.9334 |
| 0.7103 | 2.73 | 30 | 0.0051 | nan | 0.9893 | 0.0043 | 0.0 | 0.9604 | 0.6729 | 0.4972 | 0.3216 | 0.9602 |
| 0.676 | 3.64 | 40 | 0.0021 | nan | 0.9969 | 0.0020 | 0.0 | 0.9675 | 0.5336 | 0.4995 | 0.3232 | 0.9675 |
| 0.5955 | 4.55 | 50 | 0.0001 | nan | 0.9993 | 0.0001 | 0.0 | 0.9698 | 0.4440 | 0.4997 | 0.3233 | 0.9698 |
| 0.5691 | 5.45 | 60 | 0.0000 | nan | 0.9997 | 0.0000 | 0.0 | 0.9702 | 0.3812 | 0.4999 | 0.3234 | 0.9702 |
| 0.5067 | 6.36 | 70 | 0.0 | nan | 0.9996 | 0.0 | 0.0 | 0.9701 | 0.3590 | 0.4998 | 0.3234 | 0.9701 |
| 0.4656 | 7.27 | 80 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9703 | 0.3247 | 0.4999 | 0.3234 | 0.9703 |
| 0.4227 | 8.18 | 90 | 0.0 | nan | 0.9998 | 0.0 | 0.0 | 0.9702 | 0.3171 | 0.4999 | 0.3234 | 0.9702 |
| 0.3898 | 9.09 | 100 | 0.0004 | nan | 0.9996 | 0.0004 | 0.0 | 0.9701 | 0.3122 | 0.5000 | 0.3235 | 0.9701 |
| 0.3513 | 10.0 | 110 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9703 | 0.2876 | 0.4999 | 0.3234 | 0.9703 |
| 0.4157 | 10.91 | 120 | 0.0000 | nan | 0.9998 | 0.0000 | 0.0 | 0.9703 | 0.2820 | 0.4999 | 0.3234 | 0.9703 |
| 0.3317 | 11.82 | 130 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9703 | 0.2693 | 0.4999 | 0.3234 | 0.9703 |
| 0.321 | 12.73 | 140 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9704 | 0.2647 | 0.4999 | 0.3235 | 0.9704 |
| 0.2887 | 13.64 | 150 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9704 | 0.2539 | 0.5000 | 0.3235 | 0.9704 |
| 0.3008 | 14.55 | 160 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9704 | 0.2536 | 0.5000 | 0.3235 | 0.9704 |
| 0.2853 | 15.45 | 170 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9704 | 0.2397 | 0.5000 | 0.3235 | 0.9704 |
| 0.2684 | 16.36 | 180 | 0.0 | nan | 0.9999 | 0.0 | 0.0 | 0.9704 | 0.2321 | 0.5000 | 0.3235 | 0.9704 |
| 0.2585 | 17.27 | 190 | 0.0000 | nan | 0.9999 | 0.0000 | 0.0 | 0.9704 | 0.2208 | 0.5000 | 0.3235 | 0.9704 |
| 0.2088 | 18.18 | 200 | 0.0084 | nan | 0.9997 | 0.0083 | 0.0 | 0.9704 | 0.2011 | 0.5041 | 0.3262 | 0.9704 |
| 0.2518 | 19.09 | 210 | 0.0468 | nan | 0.9989 | 0.0451 | 0.0 | 0.9707 | 0.2026 | 0.5228 | 0.3386 | 0.9707 |
| 0.218 | 20.0 | 220 | 0.0879 | nan | 0.9984 | 0.0834 | nan | 0.9714 | 0.1889 | 0.5431 | 0.5274 | 0.9715 |
| 0.2046 | 20.91 | 230 | 0.1931 | nan | 0.9969 | 0.1752 | nan | 0.9730 | 0.1847 | 0.5950 | 0.5741 | 0.9732 |
| 0.2147 | 21.82 | 240 | 0.2042 | nan | 0.9968 | 0.1850 | nan | 0.9733 | 0.1766 | 0.6005 | 0.5791 | 0.9734 |
| 0.188 | 22.73 | 250 | 0.2020 | nan | 0.9972 | 0.1849 | nan | 0.9735 | 0.1726 | 0.5996 | 0.5792 | 0.9737 |
| 0.2175 | 23.64 | 260 | 0.1898 | nan | 0.9974 | 0.1748 | nan | 0.9734 | 0.1706 | 0.5936 | 0.5741 | 0.9735 |
| 0.2059 | 24.55 | 270 | 0.3006 | nan | 0.9962 | 0.2670 | nan | 0.9754 | 0.1689 | 0.6484 | 0.6212 | 0.9756 |
| 0.1776 | 25.45 | 280 | 0.2870 | nan | 0.9967 | 0.2587 | nan | 0.9755 | 0.1612 | 0.6418 | 0.6171 | 0.9757 |
| 0.1585 | 26.36 | 290 | 0.4254 | nan | 0.9944 | 0.3593 | nan | 0.9773 | 0.1537 | 0.7099 | 0.6683 | 0.9776 |
| 0.1588 | 27.27 | 300 | 0.2798 | nan | 0.9970 | 0.2548 | nan | 0.9756 | 0.1527 | 0.6384 | 0.6152 | 0.9758 |
| 0.153 | 28.18 | 310 | 0.4288 | nan | 0.9946 | 0.3646 | nan | 0.9776 | 0.1452 | 0.7117 | 0.6711 | 0.9779 |
| 0.1623 | 29.09 | 320 | 0.4401 | nan | 0.9945 | 0.3726 | nan | 0.9778 | 0.1442 | 0.7173 | 0.6752 | 0.9781 |
| 0.1603 | 30.0 | 330 | 0.4050 | nan | 0.9958 | 0.3562 | nan | 0.9781 | 0.1407 | 0.7004 | 0.6671 | 0.9784 |
| 0.1694 | 30.91 | 340 | 0.4585 | nan | 0.9948 | 0.3911 | nan | 0.9786 | 0.1343 | 0.7266 | 0.6849 | 0.9789 |
| 0.1585 | 31.82 | 350 | 0.3861 | nan | 0.9962 | 0.3433 | nan | 0.9779 | 0.1353 | 0.6912 | 0.6606 | 0.9782 |
| 0.1342 | 32.73 | 360 | 0.4963 | nan | 0.9939 | 0.4132 | nan | 0.9789 | 0.1338 | 0.7451 | 0.6961 | 0.9792 |
| 0.1358 | 33.64 | 370 | 0.5048 | nan | 0.9937 | 0.4182 | nan | 0.9789 | 0.1342 | 0.7493 | 0.6986 | 0.9793 |
| 0.1493 | 34.55 | 380 | 0.4809 | nan | 0.9946 | 0.4080 | nan | 0.9791 | 0.1297 | 0.7377 | 0.6936 | 0.9794 |
| 0.1435 | 35.45 | 390 | 0.5658 | nan | 0.9923 | 0.4518 | nan | 0.9794 | 0.1271 | 0.7791 | 0.7156 | 0.9797 |
| 0.1305 | 36.36 | 400 | 0.4157 | nan | 0.9968 | 0.3758 | nan | 0.9793 | 0.1225 | 0.7062 | 0.6776 | 0.9796 |
| 0.1496 | 37.27 | 410 | 0.5385 | nan | 0.9934 | 0.4420 | nan | 0.9796 | 0.1237 | 0.7659 | 0.7108 | 0.9799 |
| 0.1445 | 38.18 | 420 | 0.5763 | nan | 0.9924 | 0.4615 | nan | 0.9798 | 0.1207 | 0.7843 | 0.7206 | 0.9801 |
| 0.1307 | 39.09 | 430 | 0.4853 | nan | 0.9956 | 0.4244 | nan | 0.9803 | 0.1194 | 0.7404 | 0.7023 | 0.9806 |
| 0.1379 | 40.0 | 440 | 0.5722 | nan | 0.9922 | 0.4557 | nan | 0.9795 | 0.1174 | 0.7822 | 0.7176 | 0.9798 |
| 0.1202 | 40.91 | 450 | 0.5399 | nan | 0.9943 | 0.4544 | nan | 0.9805 | 0.1143 | 0.7671 | 0.7175 | 0.9809 |
| 0.1239 | 41.82 | 460 | 0.5580 | nan | 0.9932 | 0.4558 | nan | 0.9800 | 0.1150 | 0.7756 | 0.7179 | 0.9803 |
| 0.1183 | 42.73 | 470 | 0.4777 | nan | 0.9961 | 0.4236 | nan | 0.9805 | 0.1129 | 0.7369 | 0.7021 | 0.9808 |
| 0.1202 | 43.64 | 480 | 0.5933 | nan | 0.9928 | 0.4793 | nan | 0.9806 | 0.1119 | 0.7930 | 0.7300 | 0.9810 |
| 0.1276 | 44.55 | 490 | 0.5425 | nan | 0.9942 | 0.4561 | nan | 0.9806 | 0.1131 | 0.7683 | 0.7183 | 0.9809 |
| 0.1172 | 45.45 | 500 | 0.6272 | nan | 0.9898 | 0.4700 | nan | 0.9787 | 0.1135 | 0.8085 | 0.7244 | 0.9791 |
| 0.1288 | 46.36 | 510 | 0.4236 | nan | 0.9974 | 0.3898 | nan | 0.9802 | 0.1105 | 0.7105 | 0.6850 | 0.9804 |
| 0.1185 | 47.27 | 520 | 0.6035 | nan | 0.9914 | 0.4711 | nan | 0.9796 | 0.1130 | 0.7975 | 0.7254 | 0.9800 |
| 0.1045 | 48.18 | 530 | 0.5750 | nan | 0.9930 | 0.4679 | nan | 0.9804 | 0.1102 | 0.7840 | 0.7241 | 0.9807 |
| 0.1211 | 49.09 | 540 | 0.5812 | nan | 0.9929 | 0.4715 | nan | 0.9804 | 0.1069 | 0.7870 | 0.7260 | 0.9808 |
| 0.1206 | 50.0 | 550 | 0.5221 | nan | 0.9953 | 0.4528 | nan | 0.9811 | 0.1071 | 0.7587 | 0.7169 | 0.9814 |
| 0.1193 | 50.91 | 560 | 0.4956 | nan | 0.9961 | 0.4398 | nan | 0.9811 | 0.1053 | 0.7459 | 0.7105 | 0.9814 |
| 0.1116 | 51.82 | 570 | 0.5257 | nan | 0.9951 | 0.4528 | nan | 0.9809 | 0.1043 | 0.7604 | 0.7169 | 0.9812 |
| 0.1218 | 52.73 | 580 | 0.5936 | nan | 0.9922 | 0.4724 | nan | 0.9801 | 0.1078 | 0.7929 | 0.7262 | 0.9804 |
| 0.1284 | 53.64 | 590 | 0.5872 | nan | 0.9924 | 0.4696 | nan | 0.9801 | 0.1054 | 0.7898 | 0.7248 | 0.9804 |
| 0.096 | 54.55 | 600 | 0.5451 | nan | 0.9942 | 0.4580 | nan | 0.9806 | 0.1028 | 0.7697 | 0.7193 | 0.9809 |
| 0.1091 | 55.45 | 610 | 0.6014 | nan | 0.9917 | 0.4725 | nan | 0.9798 | 0.1022 | 0.7965 | 0.7261 | 0.9802 |
| 0.1068 | 56.36 | 620 | 0.4926 | nan | 0.9962 | 0.4374 | nan | 0.9810 | 0.1015 | 0.7444 | 0.7092 | 0.9813 |
| 0.106 | 57.27 | 630 | 0.5713 | nan | 0.9937 | 0.4731 | nan | 0.9809 | 0.1011 | 0.7825 | 0.7270 | 0.9812 |
| 0.1009 | 58.18 | 640 | 0.4512 | nan | 0.9969 | 0.4089 | nan | 0.9805 | 0.1028 | 0.7240 | 0.6947 | 0.9807 |
| 0.1018 | 59.09 | 650 | 0.6053 | nan | 0.9919 | 0.4779 | nan | 0.9801 | 0.1022 | 0.7986 | 0.7290 | 0.9805 |
| 0.1012 | 60.0 | 660 | 0.5167 | nan | 0.9949 | 0.4427 | nan | 0.9805 | 0.1016 | 0.7558 | 0.7116 | 0.9808 |
| 0.1052 | 60.91 | 670 | 0.5464 | nan | 0.9943 | 0.4604 | nan | 0.9808 | 0.0999 | 0.7703 | 0.7206 | 0.9811 |
| 0.1229 | 61.82 | 680 | 0.5706 | nan | 0.9939 | 0.4750 | nan | 0.9810 | 0.0993 | 0.7822 | 0.7280 | 0.9814 |
| 0.0963 | 62.73 | 690 | 0.5746 | nan | 0.9936 | 0.4754 | nan | 0.9809 | 0.0974 | 0.7841 | 0.7282 | 0.9813 |
| 0.1115 | 63.64 | 700 | 0.5239 | nan | 0.9955 | 0.4562 | nan | 0.9813 | 0.0974 | 0.7597 | 0.7187 | 0.9816 |
| 0.1025 | 64.55 | 710 | 0.5845 | nan | 0.9935 | 0.4813 | nan | 0.9811 | 0.0964 | 0.7890 | 0.7312 | 0.9814 |
| 0.0916 | 65.45 | 720 | 0.5493 | nan | 0.9947 | 0.4685 | nan | 0.9813 | 0.0962 | 0.7720 | 0.7249 | 0.9816 |
| 0.1055 | 66.36 | 730 | 0.5273 | nan | 0.9953 | 0.4571 | nan | 0.9812 | 0.0947 | 0.7613 | 0.7191 | 0.9815 |
| 0.1081 | 67.27 | 740 | 0.6093 | nan | 0.9919 | 0.4813 | nan | 0.9802 | 0.0964 | 0.8006 | 0.7308 | 0.9806 |
| 0.1039 | 68.18 | 750 | 0.5405 | nan | 0.9945 | 0.4573 | nan | 0.9807 | 0.0950 | 0.7675 | 0.7190 | 0.9811 |
| 0.106 | 69.09 | 760 | 0.5564 | nan | 0.9943 | 0.4682 | nan | 0.9810 | 0.0939 | 0.7753 | 0.7246 | 0.9813 |
| 0.0912 | 70.0 | 770 | 0.5377 | nan | 0.9949 | 0.4612 | nan | 0.9811 | 0.0936 | 0.7663 | 0.7212 | 0.9814 |
| 0.0951 | 70.91 | 780 | 0.5600 | nan | 0.9941 | 0.4689 | nan | 0.9809 | 0.0938 | 0.7771 | 0.7249 | 0.9813 |
| 0.0998 | 71.82 | 790 | 0.5573 | nan | 0.9944 | 0.4705 | nan | 0.9812 | 0.0928 | 0.7759 | 0.7258 | 0.9815 |
| 0.0889 | 72.73 | 800 | 0.5398 | nan | 0.9949 | 0.4628 | nan | 0.9812 | 0.0931 | 0.7674 | 0.7220 | 0.9815 |
| 0.0906 | 73.64 | 810 | 0.5151 | nan | 0.9958 | 0.4528 | nan | 0.9813 | 0.0928 | 0.7555 | 0.7171 | 0.9816 |
| 0.0911 | 74.55 | 820 | 0.5682 | nan | 0.9938 | 0.4722 | nan | 0.9809 | 0.0924 | 0.7810 | 0.7265 | 0.9812 |
| 0.0907 | 75.45 | 830 | 0.4864 | nan | 0.9965 | 0.4365 | nan | 0.9812 | 0.0929 | 0.7415 | 0.7089 | 0.9815 |
| 0.1117 | 76.36 | 840 | 0.5239 | nan | 0.9956 | 0.4576 | nan | 0.9814 | 0.0934 | 0.7598 | 0.7195 | 0.9817 |
| 0.0812 | 77.27 | 850 | 0.5279 | nan | 0.9956 | 0.4605 | nan | 0.9814 | 0.0915 | 0.7617 | 0.7210 | 0.9817 |
| 0.0888 | 78.18 | 860 | 0.5615 | nan | 0.9942 | 0.4720 | nan | 0.9811 | 0.0915 | 0.7778 | 0.7266 | 0.9814 |
| 0.09 | 79.09 | 870 | 0.5414 | nan | 0.9948 | 0.4628 | nan | 0.9811 | 0.0920 | 0.7681 | 0.7220 | 0.9814 |
| 0.1052 | 80.0 | 880 | 0.5866 | nan | 0.9932 | 0.4790 | nan | 0.9808 | 0.0917 | 0.7899 | 0.7299 | 0.9812 |
| 0.0867 | 80.91 | 890 | 0.5252 | nan | 0.9955 | 0.4573 | nan | 0.9813 | 0.0912 | 0.7603 | 0.7193 | 0.9816 |
| 0.0942 | 81.82 | 900 | 0.5091 | nan | 0.9959 | 0.4490 | nan | 0.9813 | 0.0925 | 0.7525 | 0.7152 | 0.9815 |
| 0.0917 | 82.73 | 910 | 0.5454 | nan | 0.9950 | 0.4682 | nan | 0.9814 | 0.0908 | 0.7702 | 0.7248 | 0.9817 |
| 0.103 | 83.64 | 920 | 0.5452 | nan | 0.9949 | 0.4672 | nan | 0.9813 | 0.0912 | 0.7701 | 0.7243 | 0.9816 |
| 0.0939 | 84.55 | 930 | 0.5539 | nan | 0.9947 | 0.4717 | nan | 0.9814 | 0.0900 | 0.7743 | 0.7265 | 0.9817 |
| 0.0892 | 85.45 | 940 | 0.5330 | nan | 0.9954 | 0.4635 | nan | 0.9815 | 0.0900 | 0.7642 | 0.7225 | 0.9818 |
| 0.0899 | 86.36 | 950 | 0.5756 | nan | 0.9938 | 0.4778 | nan | 0.9811 | 0.0905 | 0.7847 | 0.7295 | 0.9814 |
| 0.0877 | 87.27 | 960 | 0.5771 | nan | 0.9937 | 0.4787 | nan | 0.9811 | 0.0893 | 0.7854 | 0.7299 | 0.9814 |
| 0.0851 | 88.18 | 970 | 0.5087 | nan | 0.9961 | 0.4512 | nan | 0.9814 | 0.0897 | 0.7524 | 0.7163 | 0.9817 |
| 0.0857 | 89.09 | 980 | 0.5363 | nan | 0.9953 | 0.4644 | nan | 0.9814 | 0.0894 | 0.7658 | 0.7229 | 0.9817 |
| 0.0821 | 90.0 | 990 | 0.5333 | nan | 0.9953 | 0.4623 | nan | 0.9814 | 0.0895 | 0.7643 | 0.7218 | 0.9817 |
| 0.0931 | 90.91 | 1000 | 0.5581 | nan | 0.9944 | 0.4718 | nan | 0.9812 | 0.0895 | 0.7763 | 0.7265 | 0.9815 |
| 0.0787 | 91.82 | 1010 | 0.5525 | nan | 0.9946 | 0.4689 | nan | 0.9812 | 0.0889 | 0.7735 | 0.7251 | 0.9815 |
| 0.0865 | 92.73 | 1020 | 0.5659 | nan | 0.9941 | 0.4746 | nan | 0.9812 | 0.0883 | 0.7800 | 0.7279 | 0.9815 |
| 0.0939 | 93.64 | 1030 | 0.5583 | nan | 0.9945 | 0.4723 | nan | 0.9813 | 0.0891 | 0.7764 | 0.7268 | 0.9816 |
| 0.0874 | 94.55 | 1040 | 0.5258 | nan | 0.9955 | 0.4580 | nan | 0.9813 | 0.0893 | 0.7607 | 0.7197 | 0.9816 |
| 0.0927 | 95.45 | 1050 | 0.5319 | nan | 0.9953 | 0.4608 | nan | 0.9813 | 0.0894 | 0.7636 | 0.7211 | 0.9816 |
| 0.0808 | 96.36 | 1060 | 0.5444 | nan | 0.9949 | 0.4665 | nan | 0.9813 | 0.0897 | 0.7696 | 0.7239 | 0.9816 |
| 0.0924 | 97.27 | 1070 | 0.5445 | nan | 0.9950 | 0.4671 | nan | 0.9814 | 0.0892 | 0.7697 | 0.7243 | 0.9817 |
| 0.08 | 98.18 | 1080 | 0.5522 | nan | 0.9947 | 0.4703 | nan | 0.9813 | 0.0884 | 0.7735 | 0.7258 | 0.9816 |
| 0.0798 | 99.09 | 1090 | 0.0880 | 0.7300 | 0.7842 | 0.9815 | nan | 0.5745 | 0.9939 | nan | 0.4788 | 0.9812 |
| 0.0789 | 100.0 | 1100 | 0.0877 | 0.7231 | 0.7668 | 0.9817 | nan | 0.5383 | 0.9952 | nan | 0.4647 | 0.9814 |
| 0.0801 | 100.91 | 1110 | 0.0885 | 0.7232 | 0.7677 | 0.9816 | nan | 0.5404 | 0.9951 | nan | 0.4650 | 0.9813 |
| 0.1043 | 101.82 | 1120 | 0.0891 | 0.7242 | 0.7697 | 0.9816 | nan | 0.5445 | 0.9950 | nan | 0.4670 | 0.9813 |
| 0.0893 | 102.73 | 1130 | 0.0882 | 0.7263 | 0.7728 | 0.9817 | nan | 0.5508 | 0.9949 | nan | 0.4712 | 0.9814 |
| 0.0923 | 103.64 | 1140 | 0.0892 | 0.7134 | 0.7504 | 0.9815 | nan | 0.5048 | 0.9960 | nan | 0.4457 | 0.9812 |
| 0.0915 | 104.55 | 1150 | 0.0884 | 0.7293 | 0.7795 | 0.9817 | nan | 0.5646 | 0.9944 | nan | 0.4772 | 0.9814 |
| 0.0859 | 105.45 | 1160 | 0.0880 | 0.7340 | 0.7941 | 0.9815 | nan | 0.5949 | 0.9932 | nan | 0.4869 | 0.9811 |
| 0.0872 | 106.36 | 1170 | 0.0872 | 0.7298 | 0.7815 | 0.9817 | nan | 0.5688 | 0.9942 | nan | 0.4783 | 0.9814 |
| 0.0845 | 107.27 | 1180 | 0.0881 | 0.7310 | 0.7843 | 0.9817 | nan | 0.5746 | 0.9940 | nan | 0.4806 | 0.9813 |
| 0.0842 | 108.18 | 1190 | 0.0869 | 0.7285 | 0.7766 | 0.9818 | nan | 0.5584 | 0.9947 | nan | 0.4755 | 0.9815 |
| 0.0906 | 109.09 | 1200 | 0.0875 | 0.7277 | 0.7754 | 0.9818 | nan | 0.5560 | 0.9947 | nan | 0.4740 | 0.9815 |
| 0.0953 | 110.0 | 1210 | 0.0878 | 0.7289 | 0.7777 | 0.9818 | nan | 0.5608 | 0.9946 | nan | 0.4764 | 0.9815 |
| 0.0988 | 110.91 | 1220 | 0.0880 | 0.7303 | 0.7809 | 0.9818 | nan | 0.5674 | 0.9944 | nan | 0.4790 | 0.9815 |
| 0.0894 | 111.82 | 1230 | 0.0869 | 0.7300 | 0.7801 | 0.9818 | nan | 0.5657 | 0.9945 | nan | 0.4785 | 0.9815 |
| 0.0788 | 112.73 | 1240 | 0.0868 | 0.7283 | 0.7758 | 0.9818 | nan | 0.5569 | 0.9948 | nan | 0.4750 | 0.9815 |
| 0.0793 | 113.64 | 1250 | 0.0870 | 0.7281 | 0.7758 | 0.9818 | nan | 0.5569 | 0.9947 | nan | 0.4747 | 0.9815 |
| 0.084 | 114.55 | 1260 | 0.0874 | 0.7295 | 0.7809 | 0.9817 | nan | 0.5675 | 0.9943 | nan | 0.4777 | 0.9814 |
| 0.0832 | 115.45 | 1270 | 0.0875 | 0.7277 | 0.7760 | 0.9817 | nan | 0.5574 | 0.9946 | nan | 0.4739 | 0.9814 |
| 0.0833 | 116.36 | 1280 | 0.0873 | 0.7274 | 0.7755 | 0.9817 | nan | 0.5563 | 0.9947 | nan | 0.4735 | 0.9814 |
| 0.0786 | 117.27 | 1290 | 0.0867 | 0.7277 | 0.7754 | 0.9818 | nan | 0.5561 | 0.9947 | nan | 0.4740 | 0.9815 |
| 0.0839 | 118.18 | 1300 | 0.0865 | 0.7285 | 0.7779 | 0.9817 | nan | 0.5613 | 0.9945 | nan | 0.4755 | 0.9814 |
| 0.0847 | 119.09 | 1310 | 0.0877 | 0.7293 | 0.7816 | 0.9816 | nan | 0.5691 | 0.9941 | nan | 0.4773 | 0.9813 |
| 0.0933 | 120.0 | 1320 | 0.0867 | 0.7280 | 0.7762 | 0.9818 | nan | 0.5578 | 0.9947 | nan | 0.4745 | 0.9814 |
### Framework versions
- Transformers 4.30.2
- Pytorch 2.0.1+cu117
- Datasets 2.13.1
- Tokenizers 0.13.3
| {"license": "other", "tags": ["vision", "image-segmentation", "generated_from_trainer"], "model-index": [{"name": "safety-utcustom-train-SF-RGBD-b5", "results": []}]} | image-segmentation | sam1120/safety-utcustom-train-SF-RGBD-b5 | [
"transformers",
"pytorch",
"tensorboard",
"segformer",
"vision",
"image-segmentation",
"generated_from_trainer",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-06T05:10:19+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us
| safety-utcustom-train-SF-RGBD-b5
================================
This model is a fine-tuned version of nvidia/mit-b5 on the sam1120/safety-utcustom-TRAIN dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0867
* Mean Iou: 0.7280
* Mean Accuracy: 0.7762
* Overall Accuracy: 0.9818
* Accuracy Unlabeled: nan
* Accuracy Safe: 0.5578
* Accuracy Unsafe: 0.9947
* Iou Unlabeled: nan
* Iou Safe: 0.4745
* Iou Unsafe: 0.9814
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 4e-06
* train\_batch\_size: 15
* eval\_batch\_size: 15
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.05
* num\_epochs: 120
### Training results
### Framework versions
* Transformers 4.30.2
* Pytorch 2.0.1+cu117
* Datasets 2.13.1
* Tokenizers 0.13.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-06\n* train\\_batch\\_size: 15\n* eval\\_batch\\_size: 15\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 120",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-06\n* train\\_batch\\_size: 15\n* eval\\_batch\\_size: 15\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 120",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
48,
117,
4,
33
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-06\n* train\\_batch\\_size: 15\n* eval\\_batch\\_size: 15\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 120### Training results### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
-0.10608691722154617,
0.03593092039227486,
-0.001977757550776005,
0.11570718884468079,
0.17274364829063416,
0.02880287356674671,
0.1164056807756424,
0.11417382210493088,
-0.11002407968044281,
0.03021562658250332,
0.10483107715845108,
0.14139693975448608,
0.01553534995764494,
0.09455715119838715,
-0.02047760784626007,
-0.3077249526977539,
-0.025350933894515038,
0.03178860619664192,
-0.08702671527862549,
0.12561443448066711,
0.06431467086076736,
-0.1626301258802414,
0.0913885310292244,
-0.0041993604972958565,
-0.2213001698255539,
0.016075706109404564,
-0.004638391546905041,
-0.031258247792720795,
0.1599779725074768,
0.023195939138531685,
0.11598169803619385,
0.009773077443242073,
0.11373211443424225,
-0.20232029259204865,
0.01815521903336048,
0.05520845577120781,
-0.004302386660128832,
0.06771158427000046,
0.06416109204292297,
0.002084829146042466,
0.15074364840984344,
-0.10706237703561783,
0.06794559210538864,
0.001361866481602192,
-0.14491651952266693,
-0.212060809135437,
-0.0734177678823471,
0.02312960848212242,
0.07788404077291489,
0.09719705581665039,
-0.005637998227030039,
0.11262689530849457,
-0.09264029562473297,
0.11236855387687683,
0.26985809206962585,
-0.2442440539598465,
-0.08641669899225235,
0.037921514362096786,
0.0004658354155253619,
0.06747421622276306,
-0.13404807448387146,
0.009291055612266064,
0.033014990389347076,
0.04683414101600647,
0.11740310490131378,
-0.032833803445100784,
-0.09761381149291992,
0.02702878974378109,
-0.1383134126663208,
-0.03278490528464317,
0.05566296726465225,
0.05408550426363945,
-0.02046424336731434,
-0.03268960118293762,
-0.06696383655071259,
-0.1808510571718216,
-0.06726883351802826,
0.011955734342336655,
0.06691421568393707,
-0.06060151383280754,
-0.11557432264089584,
-0.01450582779943943,
-0.11067382991313934,
-0.08459050953388214,
-0.05048304796218872,
0.12783010303974152,
0.034354519098997116,
0.018296562135219574,
-0.034879881888628006,
0.12622380256652832,
-0.02817385643720627,
-0.13920873403549194,
0.017701096832752228,
0.029305465519428253,
-0.04261651635169983,
-0.03293188288807869,
-0.04918186739087105,
-0.06305243819952011,
-0.01254442147910595,
0.108416348695755,
-0.06065036356449127,
0.06757199019193649,
0.037490494549274445,
0.04950343072414398,
-0.11385311186313629,
0.18772754073143005,
-0.06702958047389984,
-0.009903568774461746,
-0.03723028302192688,
0.05792734771966934,
0.003224416868761182,
-0.022249674424529076,
-0.10548724234104156,
0.004622267093509436,
0.06978011131286621,
-0.008937460370361805,
-0.08813031762838364,
0.07057182490825653,
-0.03815837204456329,
-0.011121523566544056,
-0.001619009650312364,
-0.07541143894195557,
0.045482244342565536,
-0.0003237299097236246,
-0.08345266431570053,
-0.02920893207192421,
0.05097145214676857,
0.01475012768059969,
0.014242805540561676,
0.1651073694229126,
-0.08741670846939087,
0.06340762227773666,
-0.11225062608718872,
-0.10048466175794601,
-0.00008767974941292778,
-0.08521050214767456,
0.03684204816818237,
-0.07816305011510849,
-0.14889350533485413,
-0.008630955591797829,
0.07254698872566223,
-0.039854660630226135,
0.002075816970318556,
-0.052857961505651474,
-0.0907006561756134,
0.002828320488333702,
-0.008380766957998276,
0.16382549703121185,
-0.06459616869688034,
0.12172510474920273,
0.038304392248392105,
0.0717344880104065,
-0.06478586047887802,
0.0392916202545166,
-0.08449015766382217,
0.018958676606416702,
-0.22279983758926392,
0.04244091361761093,
-0.05058697983622551,
0.06852208077907562,
-0.060167133808135986,
-0.12284883856773376,
0.00810441467911005,
0.002183999866247177,
0.0905042365193367,
0.10707087814807892,
-0.22242222726345062,
-0.07676049321889877,
0.14913198351860046,
-0.07309137284755707,
-0.09894546121358871,
0.11164408922195435,
-0.0638454258441925,
0.013634889386594296,
0.061530619859695435,
0.1978117972612381,
0.054447345435619354,
-0.13971315324306488,
0.022939130663871765,
-0.015481461770832539,
0.04800724983215332,
-0.027052126824855804,
0.05012309178709984,
0.022360987961292267,
0.08739174902439117,
0.018598539754748344,
-0.06604540348052979,
0.0683792233467102,
-0.1244339570403099,
-0.09731727838516235,
-0.025170784443616867,
-0.0863848328590393,
0.04213723912835121,
0.08993475884199142,
0.06292638182640076,
-0.10371150076389313,
-0.07755736261606216,
0.09364711493253708,
0.07575555890798569,
-0.0689447671175003,
0.03885435685515404,
-0.06526336073875427,
0.04326045140624046,
-0.01824861951172352,
-0.03586962819099426,
-0.17518778145313263,
-0.02628282643854618,
-0.022984806448221207,
0.03086429461836815,
0.029698217287659645,
0.0236675925552845,
0.09268413484096527,
0.08863966912031174,
-0.0719456598162651,
-0.02384408935904503,
-0.06463594734668732,
0.0033123015891760588,
-0.12233610451221466,
-0.2293328493833542,
-0.04240270331501961,
-0.0075140418484807014,
0.09169197082519531,
-0.2126227617263794,
0.02506515011191368,
0.022410254925489426,
0.08824993669986725,
0.025363869965076447,
-0.030960924923419952,
-0.05304301157593727,
0.0766134038567543,
-0.009694729931652546,
-0.06512396037578583,
0.07029059529304504,
-0.005183880217373371,
-0.06734685599803925,
-0.05596558749675751,
-0.11224101483821869,
0.16426627337932587,
0.13397760689258575,
-0.14651593565940857,
-0.09311582893133163,
-0.01359105110168457,
-0.06413695961236954,
-0.03307177126407623,
-0.044175099581480026,
0.038943830877542496,
0.17875206470489502,
0.000299038365483284,
0.13256259262561798,
-0.0600387267768383,
-0.03444809839129448,
0.030126770958304405,
-0.026723412796854973,
0.028342677280306816,
0.12836842238903046,
0.1247066780924797,
-0.06253943592309952,
0.12415965646505356,
0.12393768876791,
-0.08032224327325821,
0.149248406291008,
-0.033717866986989975,
-0.08056658506393433,
-0.018987931311130524,
-0.015585609711706638,
-0.009135747328400612,
0.1773144006729126,
-0.15017400681972504,
-0.017383012920618057,
-0.004827072843909264,
0.014186468906700611,
0.014743251726031303,
-0.2501024603843689,
-0.05698255077004433,
0.039848167449235916,
-0.044327512383461,
-0.011931851506233215,
-0.023646032437682152,
-0.004895104561001062,
0.10396166145801544,
-0.006570158060640097,
-0.07460574805736542,
0.0013218804961070418,
-0.007671609055250883,
-0.04921742156147957,
0.20673243701457977,
-0.05739029869437218,
-0.11817362159490585,
-0.08848458528518677,
-0.07862059026956558,
-0.03554953634738922,
0.001964328810572624,
0.05927960202097893,
-0.1085573211312294,
-0.01842140592634678,
-0.05836477130651474,
0.017396489158272743,
0.00773574085906148,
0.03737619146704674,
-0.0011294992873445153,
-0.008642701432108879,
0.05627564713358879,
-0.09601028263568878,
-0.009704096242785454,
-0.06692133843898773,
-0.05169863626360893,
0.05476170778274536,
0.05892135947942734,
0.14857777953147888,
0.13412877917289734,
-0.025005238130688667,
0.020132213830947876,
-0.032991617918014526,
0.2601378858089447,
-0.09613046795129776,
-0.028835749253630638,
0.11757852882146835,
-0.011363915167748928,
0.055678948760032654,
0.10780814290046692,
0.08346143364906311,
-0.10834025591611862,
-0.0026590940542519093,
0.06460016220808029,
-0.050982434302568436,
-0.1546965092420578,
-0.014341188594698906,
-0.05813394486904144,
-0.031197188422083855,
0.07726877927780151,
0.027075236663222313,
-0.0021959934383630753,
0.055142857134342194,
0.049514252692461014,
0.042689546942710876,
-0.023555930703878403,
0.05029090866446495,
0.08921565860509872,
0.03269160911440849,
0.10884394496679306,
-0.04491584375500679,
-0.0672847107052803,
0.0308866985142231,
0.0019098117481917143,
0.24749353528022766,
-0.01417852658778429,
0.0957847461104393,
0.07380521297454834,
0.16148434579372406,
-0.011448616161942482,
0.04818202555179596,
-0.015343602746725082,
-0.06901425868272781,
-0.01867866888642311,
-0.04442105442285538,
-0.014355833642184734,
0.010740770027041435,
-0.05162559822201729,
0.038112856447696686,
-0.12661254405975342,
0.009809238836169243,
0.06765376031398773,
0.2520061433315277,
0.02760263904929161,
-0.31661126017570496,
-0.06570599973201752,
-0.006438055075705051,
-0.01202945876866579,
-0.008167579770088196,
0.006057623773813248,
0.15122488141059875,
-0.08193928003311157,
0.05643199011683464,
-0.0857476219534874,
0.08541464060544968,
-0.036125198006629944,
0.051779840141534805,
0.07766179740428925,
0.07349259406328201,
-0.0044111269526183605,
0.05498836562037468,
-0.2865791618824005,
0.30230218172073364,
0.0017845692345872521,
0.08529753983020782,
-0.06303023546934128,
-0.03268115594983101,
0.03422747552394867,
0.0824907049536705,
0.08759971708059311,
-0.015932418406009674,
-0.01994459331035614,
-0.2131185084581375,
-0.02101345919072628,
0.031770020723342896,
0.12939825654029846,
-0.01753067784011364,
0.10389383137226105,
-0.01007583923637867,
-0.006132479291409254,
0.07500214129686356,
-0.001029675593599677,
-0.03529633209109306,
-0.09107809513807297,
-0.025893202051520348,
-0.026142172515392303,
-0.04880508407950401,
-0.05838625133037567,
-0.10744072496891022,
-0.11721088737249374,
0.11056412011384964,
0.01882246509194374,
-0.013538222759962082,
-0.12048719078302383,
0.09882857650518417,
0.07736021280288696,
-0.07548286020755768,
0.042028117924928665,
0.031494125723838806,
0.058022137731313705,
0.03313014656305313,
-0.05733157694339752,
0.11761484295129776,
-0.059804197400808334,
-0.15814289450645447,
-0.056625716388225555,
0.08941485732793808,
0.05111934617161751,
0.05697965249419212,
-0.023976625874638557,
0.015375403687357903,
-0.01754717156291008,
-0.09274747967720032,
0.054551057517528534,
-0.04462480545043945,
0.06318151950836182,
0.008224936202168465,
-0.021059725433588028,
0.05080724135041237,
-0.05676502734422684,
-0.012788014486432076,
0.14655734598636627,
0.286176860332489,
-0.08971082419157028,
0.012418803758919239,
0.016965514048933983,
-0.06492669135332108,
-0.18953804671764374,
0.08004980534315109,
0.058572571724653244,
-0.00019354760297574103,
0.08573818951845169,
-0.1673789620399475,
0.09860655665397644,
0.10285788774490356,
0.0004287587944418192,
0.11358029395341873,
-0.3652622401714325,
-0.12765061855316162,
0.08024778217077255,
0.19131873548030853,
0.07842548191547394,
-0.15406708419322968,
0.001008150982670486,
-0.0019230941543355584,
-0.14958488941192627,
0.09183278679847717,
-0.08011125028133392,
0.13521991670131683,
-0.01961558870971203,
0.08635915815830231,
0.01583431102335453,
-0.061816368252038956,
0.12198234349489212,
-0.003130296478047967,
0.14073938131332397,
-0.06897472590208054,
-0.0410151481628418,
0.05346040055155754,
-0.03787672519683838,
-0.013352053239941597,
-0.04568612203001976,
0.026305589824914932,
-0.05873160436749458,
-0.011890747584402561,
-0.10534968227148056,
0.013024079613387585,
-0.0386519692838192,
-0.06721772998571396,
-0.04488993063569069,
0.04398706927895546,
0.045481353998184204,
-0.003851782064884901,
0.15254844725131989,
-0.01130048930644989,
0.11455438286066055,
0.046330418437719345,
0.058767735958099365,
-0.0615517720580101,
-0.10718003660440445,
-0.018076620995998383,
0.008423111401498318,
0.04817817732691765,
-0.13607391715049744,
0.015327856875956059,
0.1530584841966629,
0.05044127628207207,
0.12353261560201645,
0.08613076061010361,
-0.030781202018260956,
0.031342774629592896,
0.06985030323266983,
-0.15737731754779816,
-0.1154409646987915,
0.0020083850249648094,
-0.06917870044708252,
-0.0734187588095665,
0.052085988223552704,
0.0772743672132492,
-0.07515420764684677,
0.012515818700194359,
-0.006306169088929892,
0.0049201613292098045,
-0.06791393458843231,
0.20507901906967163,
0.05563955754041672,
0.040923818945884705,
-0.10418830066919327,
0.07234484702348709,
0.01792915165424347,
-0.08870413154363632,
-0.0023048901930451393,
0.09115592390298843,
-0.07095689326524734,
-0.024960309267044067,
0.08077192306518555,
0.19374996423721313,
-0.0779162049293518,
-0.02246452122926712,
-0.15024343132972717,
-0.10667082667350769,
0.06975654512643814,
0.1859237253665924,
0.09955276548862457,
-0.00778448348864913,
-0.052652060985565186,
0.04848532751202583,
-0.11685623228549957,
0.07737495750188828,
0.024113591760396957,
0.08102607727050781,
-0.14975398778915405,
0.18171575665473938,
0.01253711897879839,
0.05438702553510666,
-0.026170259341597557,
0.03222624212503433,
-0.11912485957145691,
0.039890531450510025,
-0.11481824517250061,
-0.03773104026913643,
-0.01601349003612995,
0.004690362140536308,
-0.013267268426716328,
-0.06220868602395058,
-0.06380006670951843,
0.004870776552706957,
-0.1284666657447815,
-0.02160790003836155,
0.046656284481287,
0.023563358932733536,
-0.1257108896970749,
-0.03826872631907463,
0.026626378297805786,
-0.06392906606197357,
0.056153539568185806,
0.03750944882631302,
0.014420386403799057,
0.06600870192050934,
-0.1710670292377472,
-0.022041039541363716,
0.06946801394224167,
-0.008001652546226978,
0.06401456892490387,
-0.03728702664375305,
-0.02623170241713524,
-0.029274269938468933,
0.08805812895298004,
0.012874494306743145,
0.061339713633060455,
-0.1364975869655609,
0.0071281809359788895,
-0.03178101405501366,
-0.09222909808158875,
-0.05959104001522064,
0.052249204367399216,
0.06305132061243057,
0.035943999886512756,
0.16357554495334625,
-0.08276907354593277,
0.04474888741970062,
-0.2178874909877777,
-0.016095662489533424,
0.002108370652422309,
-0.10795087367296219,
-0.08196807652711868,
-0.07301173359155655,
0.08238652348518372,
-0.07628366351127625,
0.11037002503871918,
0.03603847697377205,
0.06559242308139801,
0.03174719586968422,
-0.03256651759147644,
-0.0049418858252465725,
0.03420382738113403,
0.2113945037126541,
0.012792229652404785,
-0.03304234519600868,
0.08932989090681076,
0.08009646087884903,
0.09972316771745682,
0.1367981880903244,
0.22665897011756897,
0.15465623140335083,
-0.025058912113308907,
0.08887926489114761,
0.051674988120794296,
-0.06469795852899551,
-0.1721952259540558,
0.03922383859753609,
-0.0527772419154644,
0.09886492043733597,
-0.062011733651161194,
0.20217464864253998,
0.08833008259534836,
-0.1829918622970581,
0.06643955409526825,
-0.04590704292058945,
-0.10109695047140121,
-0.08048862218856812,
-0.03577698767185211,
-0.06985839456319809,
-0.14901278913021088,
0.026417793706059456,
-0.10164396464824677,
0.04281714931130409,
0.15033011138439178,
0.010645571164786816,
-0.011791165918111801,
0.21398882567882538,
0.035106223076581955,
0.03552377223968506,
0.05790330842137337,
0.01330779679119587,
-0.030400004237890244,
-0.09301112592220306,
-0.05913887172937393,
0.017870938405394554,
-0.03202761337161064,
0.017093077301979065,
-0.06955426931381226,
-0.07782453298568726,
0.026705512776970863,
0.005225952249020338,
-0.09349881112575531,
0.023712756112217903,
0.020987264811992645,
0.08964364975690842,
0.026511840522289276,
0.005571325775235891,
0.01645137555897236,
-0.02851756475865841,
0.24700136482715607,
-0.09348006546497345,
-0.0834280252456665,
-0.08329980820417404,
0.21910478174686432,
0.0311441607773304,
0.0013993323082104325,
0.008266991004347801,
-0.08046311885118484,
0.011112791486084461,
0.23108461499214172,
0.17191335558891296,
-0.1320185661315918,
-0.009713242761790752,
0.0010651779593899846,
0.0012678061611950397,
-0.029434990137815475,
0.11686662584543228,
0.12055034190416336,
0.05160887911915779,
-0.11452575773000717,
-0.0510820634663105,
-0.053031668066978455,
-0.018810994923114777,
-0.025718800723552704,
0.05176597088575363,
0.06786521524190903,
0.02226211130619049,
-0.06912929564714432,
0.07645877450704575,
-0.05781887471675873,
-0.1422923505306244,
0.10483518987894058,
-0.22610412538051605,
-0.15584276616573334,
-0.006965226959437132,
0.12168720364570618,
0.0047918884083628654,
0.059895098209381104,
-0.041290298104286194,
0.00035296467831358314,
0.051490914076566696,
-0.004996469710022211,
-0.07831092923879623,
-0.10564941167831421,
0.08639929443597794,
-0.11585104465484619,
0.21638692915439606,
-0.05945666506886482,
0.034251146018505096,
0.11385010927915573,
0.06392377614974976,
-0.04966441169381142,
0.05567779764533043,
0.04139083996415138,
-0.12435201555490494,
-0.005178611725568771,
0.12417864054441452,
-0.03876596689224243,
0.05394002050161362,
0.03193139284849167,
-0.13062697649002075,
0.033084433525800705,
-0.05591070279479027,
-0.04176372289657593,
-0.0278988815844059,
-0.05112803354859352,
-0.062290407717227936,
0.11540388315916061,
0.2082674205303192,
-0.007841113023459911,
0.025153297930955887,
-0.08729343116283417,
0.0162569098174572,
0.0663943886756897,
0.0478595606982708,
-0.07936103641986847,
-0.21624435484409332,
0.00717351958155632,
0.0712875947356224,
-0.040728501975536346,
-0.20447930693626404,
-0.11243564635515213,
0.03727134317159653,
-0.05401773378252983,
-0.07062101364135742,
0.0895499512553215,
0.09207607805728912,
0.05662762373685837,
-0.054878562688827515,
-0.10490445792675018,
-0.05846193805336952,
0.1701471507549286,
-0.14647188782691956,
-0.07746637612581253
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | shidowake/cyber2-7B-base-bnb-4bit | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-06T05:11:07+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.049007222056388855,
0.16460949182510376,
-0.005271392408758402,
0.021910345181822777,
0.09685911983251572,
0.01403510570526123,
0.07018975168466568,
0.11002060770988464,
-0.02425350993871689,
0.11399492621421814,
0.03344893455505371,
0.09780009090900421,
0.11368958652019501,
0.1498505026102066,
-0.002398149576038122,
-0.23227156698703766,
0.04924226179718971,
-0.1249755248427391,
-0.03746527433395386,
0.1159956082701683,
0.15001481771469116,
-0.10170940309762955,
0.07611104100942612,
-0.029819702729582787,
-0.008722295984625816,
-0.032589927315711975,
-0.056551046669483185,
-0.04997202008962631,
0.051094699651002884,
0.07382578402757645,
0.06793182343244553,
0.004094683099538088,
0.09450557827949524,
-0.2669448256492615,
0.0197003111243248,
0.0730973482131958,
-0.002068581758067012,
0.07547242939472198,
0.054895199835300446,
-0.07525460422039032,
0.09282654523849487,
-0.0507965162396431,
0.1469351053237915,
0.08020289987325668,
-0.09152709692716599,
-0.19188682734966278,
-0.0887833908200264,
0.10164182633161545,
0.18469172716140747,
0.045696184039115906,
-0.022488808259367943,
0.09940612316131592,
-0.08621317893266678,
0.011039474047720432,
0.05154034495353699,
-0.06937182694673538,
-0.05223534256219864,
0.06355299055576324,
0.08018788695335388,
0.07678371667861938,
-0.12301702797412872,
-0.02094447799026966,
0.008637533523142338,
0.00831096712499857,
0.08201737701892853,
0.023290244862437248,
0.1510206013917923,
0.03883988782763481,
-0.12744688987731934,
-0.050009194761514664,
0.10665731877088547,
0.041741468012332916,
-0.04784774035215378,
-0.25138479471206665,
-0.030326439067721367,
-0.027732934802770615,
-0.029999805614352226,
-0.03873695060610771,
0.04263332113623619,
-0.0072723389603197575,
0.0826614573597908,
-0.008116158656775951,
-0.07679495960474014,
-0.03798604756593704,
0.06191713735461235,
0.060809630900621414,
0.026244111359119415,
-0.011753023602068424,
0.010934822261333466,
0.1174238994717598,
0.10631082952022552,
-0.12367359548807144,
-0.051516905426979065,
-0.06431761384010315,
-0.07867198437452316,
-0.04216236248612404,
0.03455616533756256,
0.041060756891965866,
0.049376390874385834,
0.2486443817615509,
0.017620395869016647,
0.05382118001580238,
0.03803925961256027,
0.010167144238948822,
0.06406087428331375,
0.11435336619615555,
-0.061582546681165695,
-0.09715550392866135,
-0.025186026468873024,
0.08966731280088425,
0.01176387071609497,
-0.04024789482355118,
-0.05783011019229889,
0.06293477863073349,
0.016524890437722206,
0.1202789843082428,
0.09223750233650208,
0.003793274285271764,
-0.07138240337371826,
-0.06413803994655609,
0.1937950700521469,
-0.1626761257648468,
0.04747059941291809,
0.034180231392383575,
-0.038511235266923904,
-0.0016249394975602627,
0.008853171020746231,
0.024325255304574966,
-0.021725021302700043,
0.08937039971351624,
-0.05618007108569145,
-0.041590798646211624,
-0.10983981937170029,
-0.035744234919548035,
0.03192625194787979,
0.009910091757774353,
-0.03217151761054993,
-0.031847331672906876,
-0.08444786816835403,
-0.06831640005111694,
0.09424425661563873,
-0.07356466352939606,
-0.053753651678562164,
-0.016938211396336555,
-0.07437273859977722,
0.024786023423075676,
0.01960081420838833,
0.07747352123260498,
-0.02004585787653923,
0.042900070548057556,
-0.05549933388829231,
0.06014169380068779,
0.10937028378248215,
0.033117540180683136,
-0.05445994809269905,
0.0621645413339138,
-0.2418462336063385,
0.0997670441865921,
-0.06829129904508591,
0.05325306951999664,
-0.15072302520275116,
-0.02465333603322506,
0.04913770779967308,
0.008168290369212627,
-0.010590006597340107,
0.13754788041114807,
-0.21924975514411926,
-0.027699807658791542,
0.1631394773721695,
-0.09464818984270096,
-0.07676627486944199,
0.05986984074115753,
-0.052457790821790695,
0.10692904144525528,
0.04047565534710884,
-0.026259733363986015,
0.06162377819418907,
-0.13397987186908722,
0.0005626814090646803,
-0.045883387327194214,
-0.01928110048174858,
0.15731419622898102,
0.07587230950593948,
-0.06994020938873291,
0.07348526269197464,
0.023750323802232742,
-0.023168303072452545,
-0.046913031488657,
-0.017583578824996948,
-0.1088033989071846,
0.010729904286563396,
-0.061985816806554794,
0.01937699131667614,
-0.025795195251703262,
-0.09332547336816788,
-0.028493179008364677,
-0.17521639168262482,
-0.020266273990273476,
0.08516935259103775,
-0.009352635592222214,
-0.01925206556916237,
-0.11787936836481094,
0.015734510496258736,
0.03501737862825394,
0.002549536293372512,
-0.1319509893655777,
-0.05043373629450798,
0.02751830592751503,
-0.16075198352336884,
0.033688947558403015,
-0.05403051897883415,
0.0491553395986557,
0.03133281692862511,
-0.031412381678819656,
-0.028679344803094864,
0.022094380110502243,
0.004997676704078913,
-0.014611656777560711,
-0.24550160765647888,
-0.026604164391756058,
-0.02145342156291008,
0.16796952486038208,
-0.21640902757644653,
0.0374150350689888,
0.07194960117340088,
0.15254895389080048,
0.008589224889874458,
-0.038006994873285294,
0.002335198922082782,
-0.075041763484478,
-0.03255171701312065,
-0.06050482019782066,
-0.009038056246936321,
-0.03572068363428116,
-0.05482286959886551,
0.04863523691892624,
-0.16824471950531006,
-0.029467429965734482,
0.1015508770942688,
0.06473538279533386,
-0.13604550063610077,
-0.019663551822304726,
-0.03585261106491089,
-0.042308371514081955,
-0.05517838895320892,
-0.05935737490653992,
0.10260266810655594,
0.05827045813202858,
0.04566904529929161,
-0.06485172361135483,
-0.0747392401099205,
0.0017082487465813756,
-0.019673427566885948,
-0.022536588832736015,
0.09213293343782425,
0.07581926137208939,
-0.12331884354352951,
0.09213830530643463,
0.10402927547693253,
0.08686267584562302,
0.0966128259897232,
-0.023164015263319016,
-0.08361977338790894,
-0.049845483154058456,
0.02228725142776966,
0.017598064616322517,
0.13447505235671997,
-0.007804518099874258,
0.05406574159860611,
0.04160919412970543,
-0.013909573666751385,
0.009752067737281322,
-0.09242741018533707,
0.032518286257982254,
0.03427431732416153,
-0.01857241988182068,
0.041615914553403854,
-0.039849672466516495,
0.019975949078798294,
0.09018522500991821,
0.046917494386434555,
0.04021155461668968,
0.014107138849794865,
-0.04660527780652046,
-0.11187547445297241,
0.16612006723880768,
-0.12780359387397766,
-0.23512837290763855,
-0.1463187336921692,
0.0034277087543159723,
0.03630480915307999,
-0.009390040300786495,
0.0017278295708820224,
-0.06397698074579239,
-0.11876852810382843,
-0.09194197505712509,
0.010153552517294884,
0.04896695911884308,
-0.0851091742515564,
-0.0603698305785656,
0.05686335638165474,
0.04057794436812401,
-0.14546048641204834,
0.019262617453932762,
0.04933769255876541,
-0.09224124997854233,
-0.009894786402583122,
0.08289197087287903,
0.06857553124427795,
0.18091025948524475,
0.013082148507237434,
-0.02271466888487339,
0.03428078070282936,
0.21755947172641754,
-0.13586747646331787,
0.11420658230781555,
0.1426045000553131,
-0.09194567799568176,
0.08309654146432877,
0.19839057326316833,
0.04078111797571182,
-0.10157861560583115,
0.032499175518751144,
0.018653791397809982,
-0.030491048470139503,
-0.24355553090572357,
-0.07171683013439178,
0.00034942623460665345,
-0.057900771498680115,
0.07530075311660767,
0.09018687158823013,
0.09155713021755219,
0.01583298109471798,
-0.0946493074297905,
-0.07830986380577087,
0.05305508151650429,
0.10324970632791519,
0.020061472430825233,
-0.013236436992883682,
0.09051742404699326,
-0.03375976160168648,
0.017617853358387947,
0.09066354483366013,
0.0011531224008649588,
0.17065346240997314,
0.05820678174495697,
0.18275249004364014,
0.07604338973760605,
0.07338658720254898,
0.01378361415117979,
0.01180104911327362,
0.019032908603549004,
0.02708563208580017,
-0.004741039127111435,
-0.08538748323917389,
-0.01599922962486744,
0.12008915096521378,
0.07424698024988174,
0.015674617141485214,
0.014355434104800224,
-0.04089333862066269,
0.08203015476465225,
0.17435193061828613,
-0.001506963511928916,
-0.1824604868888855,
-0.06271602213382721,
0.08220411837100983,
-0.09449198096990585,
-0.10147359222173691,
-0.02445729449391365,
0.03089604340493679,
-0.17088350653648376,
0.023070847615599632,
-0.016430631279945374,
0.11182350665330887,
-0.13931094110012054,
-0.019696295261383057,
0.0640200525522232,
0.07118809968233109,
-0.00031885437783785164,
0.05944213643670082,
-0.16128569841384888,
0.10404066741466522,
0.013166810385882854,
0.06712377816438675,
-0.09715772420167923,
0.10046469420194626,
-0.006883090827614069,
-0.013416164554655552,
0.13275203108787537,
0.008256223052740097,
-0.07161599397659302,
-0.07921489328145981,
-0.09379399567842484,
-0.009093280881643295,
0.12668752670288086,
-0.14835532009601593,
0.08585991710424423,
-0.035368360579013824,
-0.04256736859679222,
0.0022144275717437267,
-0.10755012929439545,
-0.12217973172664642,
-0.1874755620956421,
0.05520224943757057,
-0.1321607530117035,
0.039849888533353806,
-0.10649667680263519,
-0.03462952747941017,
-0.029491933062672615,
0.1882491409778595,
-0.22971367835998535,
-0.06835493445396423,
-0.15157760679721832,
-0.09785088151693344,
0.14553189277648926,
-0.04969761520624161,
0.08694402873516083,
-0.005991519894450903,
0.18016821146011353,
0.022223925217986107,
-0.021585633978247643,
0.09859558939933777,
-0.09382225573062897,
-0.1963716447353363,
-0.08180448412895203,
0.15751656889915466,
0.13459575176239014,
0.03521031513810158,
-0.0027760460507124662,
0.037876322865486145,
-0.01856307126581669,
-0.12259240448474884,
0.021658578887581825,
0.17797763645648956,
0.0652514174580574,
0.02310643345117569,
-0.026529761031270027,
-0.11104881763458252,
-0.06772379577159882,
-0.033685971051454544,
0.03064778819680214,
0.18449479341506958,
-0.0722544714808464,
0.18419069051742554,
0.143813356757164,
-0.05867353826761246,
-0.1976030021905899,
0.008879725821316242,
0.03365374729037285,
0.007196295075118542,
0.03445420414209366,
-0.20255140960216522,
0.0841677114367485,
0.00034181843511760235,
-0.05190233513712883,
0.13343381881713867,
-0.17106693983078003,
-0.15042030811309814,
0.07339101284742355,
0.03619921952486038,
-0.19460853934288025,
-0.11963265389204025,
-0.08913769572973251,
-0.05391303077340126,
-0.18051348626613617,
0.10290905088186264,
0.03496568650007248,
0.008035079576075077,
0.03376363217830658,
0.028494013473391533,
0.01669638603925705,
-0.03928735852241516,
0.1920013129711151,
-0.026591487228870392,
0.029855716973543167,
-0.08456290513277054,
-0.06990274786949158,
0.04655740037560463,
-0.05482156574726105,
0.0760476216673851,
-0.027013001963496208,
0.011612839996814728,
-0.10561433434486389,
-0.042526841163635254,
-0.029051896184682846,
0.013453613966703415,
-0.0963861495256424,
-0.08940120041370392,
-0.0490599125623703,
0.09310506284236908,
0.09519506990909576,
-0.035876575857400894,
-0.03684677556157112,
-0.07069114595651627,
0.039579302072525024,
0.18676936626434326,
0.17657315731048584,
0.04523694887757301,
-0.0789421945810318,
-0.005537794437259436,
-0.011924253776669502,
0.04352729767560959,
-0.21637341380119324,
0.06442029029130936,
0.05013522133231163,
0.017847778275609016,
0.11767403781414032,
-0.02045002020895481,
-0.1556767225265503,
-0.07006701827049255,
0.06328949332237244,
-0.06132598593831062,
-0.1951322853565216,
0.005576360039412975,
0.054395273327827454,
-0.16848263144493103,
-0.048018258064985275,
0.04364382475614548,
-0.004054433200508356,
-0.0402018167078495,
0.01867259293794632,
0.08977478742599487,
0.003425614908337593,
0.0704059898853302,
0.05869606137275696,
0.08224445581436157,
-0.10246741771697998,
0.07471306622028351,
0.08622124791145325,
-0.07954994589090347,
0.026619622483849525,
0.09149482846260071,
-0.05819176882505417,
-0.02969011478126049,
0.02704544924199581,
0.0793747529387474,
0.011502381414175034,
-0.042540501803159714,
0.011518802493810654,
-0.10228829830884933,
0.06203006953001022,
0.08760257810354233,
0.03265642002224922,
0.015443529933691025,
0.03219176456332207,
0.045628782361745834,
-0.07176384329795837,
0.1219232901930809,
0.028246978297829628,
0.015991143882274628,
-0.04067446291446686,
-0.04898078367114067,
0.024271609261631966,
-0.0303955040872097,
-0.006366716232150793,
-0.03475780412554741,
-0.0729878842830658,
-0.0171539094299078,
-0.16714228689670563,
-0.016664555296301842,
-0.04662061110138893,
0.009329318068921566,
0.03086909092962742,
-0.03788549080491066,
0.008464637212455273,
0.007407912518829107,
-0.07459274679422379,
-0.06477426737546921,
-0.022905457764863968,
0.09289900958538055,
-0.16393527388572693,
0.02335011027753353,
0.08690579235553741,
-0.12064014375209808,
0.09392421692609787,
0.01837589405477047,
-0.0037578048650175333,
0.028480252251029015,
-0.14924435317516327,
0.038928523659706116,
-0.03113253228366375,
0.014821149408817291,
0.04454975947737694,
-0.2236335128545761,
0.0009650349384173751,
-0.033828526735305786,
-0.06339430809020996,
-0.009390673600137234,
-0.036760155111551285,
-0.11370383948087692,
0.10629112273454666,
0.007970798760652542,
-0.08916810154914856,
-0.031690530478954315,
0.032128699123859406,
0.08206479996442795,
-0.0239556971937418,
0.15763959288597107,
-0.0023972811177372932,
0.0736590027809143,
-0.1675432026386261,
-0.019303109496831894,
-0.011248460970818996,
0.020926566794514656,
-0.018098697066307068,
-0.01251189224421978,
0.04078914225101471,
-0.02225574664771557,
0.18437865376472473,
-0.023570427671074867,
0.023348741233348846,
0.06592654436826706,
0.027775658294558525,
-0.025002485141158104,
0.10530006885528564,
0.05339968949556351,
0.021854043006896973,
0.02036798559129238,
0.00273964018560946,
-0.04241073876619339,
-0.023610878735780716,
-0.1998770385980606,
0.06446972489356995,
0.14037446677684784,
0.09086652100086212,
-0.017234215512871742,
0.08257289230823517,
-0.1004219725728035,
-0.11521948128938675,
0.11568495631217957,
-0.05446505919098854,
-0.004037478007376194,
-0.0672159418463707,
0.12938179075717926,
0.1446845531463623,
-0.19097456336021423,
0.06995914876461029,
-0.06848131865262985,
-0.049033988267183304,
-0.11654651165008545,
-0.1963350623846054,
-0.05714293569326401,
-0.05161691829562187,
-0.01663723587989807,
-0.046969223767519,
0.07560921460390091,
0.05719533935189247,
0.007424132898449898,
-0.0017566849710419774,
0.06332923471927643,
-0.026077456772327423,
0.00009585227962816134,
0.026813751086592674,
0.06610306352376938,
0.013093758374452591,
-0.02985633723437786,
0.017491595819592476,
-0.012147722765803337,
0.042048826813697815,
0.06357792019844055,
0.04670548066496849,
-0.030032360926270485,
0.016853880137205124,
-0.03863191977143288,
-0.10680584609508514,
0.041318636387586594,
-0.028504958376288414,
-0.08043242245912552,
0.1491626501083374,
0.02454165369272232,
0.008750278502702713,
-0.0205967016518116,
0.2416755110025406,
-0.0737907737493515,
-0.09567341208457947,
-0.1479424238204956,
0.10524045675992966,
-0.04420987144112587,
0.06244929879903793,
0.045180387794971466,
-0.10425344854593277,
0.016717668622732162,
0.12817999720573425,
0.16302813589572906,
-0.044200748205184937,
0.020526019856333733,
0.027614353224635124,
0.004152800887823105,
-0.03678637370467186,
0.0514480359852314,
0.06988705694675446,
0.1595088243484497,
-0.048713311553001404,
0.09546878933906555,
-0.0016016386216506362,
-0.09618084132671356,
-0.03802286460995674,
0.11709540337324142,
-0.018092934042215347,
0.017691975459456444,
-0.055210161954164505,
0.11857418715953827,
-0.06138255074620247,
-0.2316483110189438,
0.06108921393752098,
-0.06591550260782242,
-0.13765475153923035,
-0.02143050730228424,
0.08041442185640335,
-0.013238796964287758,
0.02708347514271736,
0.07207029312849045,
-0.07533451914787292,
0.20003929734230042,
0.037636954337358475,
-0.05420409142971039,
-0.05360380560159683,
0.08255447447299957,
-0.10376271605491638,
0.27565470337867737,
0.016520937904715538,
0.04948882386088371,
0.10317612439393997,
-0.012690499424934387,
-0.13475549221038818,
0.02108365297317505,
0.09600389003753662,
-0.0946137085556984,
0.04216265305876732,
0.19903649389743805,
0.0003853837260976434,
0.1207512691617012,
0.0790785402059555,
-0.07618726044893265,
0.049590613692998886,
-0.0941753089427948,
-0.07070460170507431,
-0.09001081436872482,
0.09455035626888275,
-0.07685617357492447,
0.14261877536773682,
0.1292559802532196,
-0.053739987313747406,
0.010677514597773552,
-0.028576120734214783,
0.04638256877660751,
0.0034859003499150276,
0.1005801111459732,
0.010024284943938255,
-0.18460705876350403,
0.02157641015946865,
0.01203901320695877,
0.1056026741862297,
-0.16518552601337433,
-0.09804878383874893,
0.042120642960071564,
0.0014211505185812712,
-0.060778699815273285,
0.12909291684627533,
0.06027422100305557,
0.04478219151496887,
-0.04292554408311844,
-0.020403601229190826,
-0.009860116057097912,
0.13677826523780823,
-0.10241927951574326,
0.0014122816501185298
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Data Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "meta-llama/Llama-2-7b-hf"} | null | simonycl/llama-2-7b-hf-cohere-KMenasRandomDeita-0.05-Llama-2-7b-hf-2e-5 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-7b-hf",
"region:us"
] | 2024-02-06T05:15:55+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
41,
6,
3,
45,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.10578131675720215,
0.17948102951049805,
-0.0035145513247698545,
0.03891412168741226,
0.08621077984571457,
0.021764913573861122,
0.05196864902973175,
0.1246534213423729,
-0.04679914936423302,
0.10144586861133575,
0.05955370143055916,
0.10162605345249176,
0.09514814615249634,
0.19481638073921204,
-0.003711341181769967,
-0.19174233078956604,
0.019053086638450623,
-0.09994730353355408,
-0.005330710206180811,
0.12174459546804428,
0.15323303639888763,
-0.09489279985427856,
0.08602795749902725,
-0.017800018191337585,
-0.013100131414830685,
-0.033980559557676315,
-0.06935591250658035,
-0.04023164510726929,
0.03871799632906914,
0.05900805443525314,
0.04762444272637367,
-0.007488492410629988,
0.08426941931247711,
-0.26613879203796387,
0.01787739433348179,
0.04328127205371857,
-0.010081695392727852,
0.08668689429759979,
0.09662588685750961,
-0.05239115655422211,
0.11235129088163376,
-0.047021105885505676,
0.12763291597366333,
0.0754753053188324,
-0.07115661352872849,
-0.17680078744888306,
-0.08671265095472336,
0.0784645527601242,
0.17059318721294403,
0.08163820952177048,
-0.038548387587070465,
0.15852737426757812,
-0.11096139997243881,
0.014471856877207756,
0.04292582347989082,
-0.04215667396783829,
-0.08149929344654083,
0.04472135007381439,
0.10571891069412231,
0.05296047776937485,
-0.1364593207836151,
-0.0316958874464035,
0.028017783537507057,
0.036908578127622604,
0.08219651132822037,
0.020722191780805588,
0.1490500420331955,
0.039162229746580124,
-0.1400441974401474,
-0.02889537438750267,
0.13521221280097961,
0.043361596763134,
-0.04594216123223305,
-0.21876536309719086,
0.009401271119713783,
-0.0688539445400238,
-0.016573267057538033,
-0.0473346970975399,
0.03277067467570305,
-0.019044723361730576,
0.08590785413980484,
-0.023151051253080368,
-0.0912981778383255,
-0.036363840103149414,
0.08296481519937515,
0.05473379045724869,
0.03431430459022522,
-0.030013028532266617,
-0.003637326881289482,
0.12796398997306824,
0.05791280418634415,
-0.12604327499866486,
-0.06496736407279968,
-0.06727676838636398,
-0.05133811756968498,
-0.05336795747280121,
0.025901637971401215,
0.030546221882104874,
0.058595795184373856,
0.22094087302684784,
-0.014979392290115356,
0.04600260779261589,
0.06179821863770485,
0.013360618613660336,
0.05709648132324219,
0.08174323290586472,
-0.07354924827814102,
-0.13286344707012177,
-0.016805140301585197,
0.09281699359416962,
-0.00784864742308855,
-0.014999274164438248,
-0.0377439446747303,
0.03574961796402931,
0.04972270876169205,
0.09172888100147247,
0.09294656664133072,
-0.008743898943066597,
-0.07654924690723419,
-0.05417617782950401,
0.20611271262168884,
-0.15715430676937103,
0.03286880999803543,
0.013141892850399017,
-0.03391672670841217,
-0.04431325942277908,
0.012842860072851181,
0.004780519288033247,
-0.02774704247713089,
0.08469497412443161,
-0.0729280412197113,
-0.025163959711790085,
-0.11791864037513733,
-0.007677200250327587,
0.03847067430615425,
0.03413010761141777,
-0.01623675785958767,
-0.02133348025381565,
-0.06823516637086868,
-0.08714993298053741,
0.09384068101644516,
-0.08491978794336319,
-0.06074149161577225,
-0.027955830097198486,
-0.09669972956180573,
0.019481539726257324,
0.013203713111579418,
0.11824406683444977,
-0.027103915810585022,
0.04453331604599953,
-0.01077277772128582,
0.04846559092402458,
0.06493427604436874,
0.03705217316746712,
-0.05129680410027504,
0.05861848220229149,
-0.1909576654434204,
0.09382237493991852,
-0.08735471218824387,
0.019278889521956444,
-0.14763744175434113,
-0.017486268654465675,
0.02822943404316902,
0.011112013831734657,
0.03018232434988022,
0.13953334093093872,
-0.21594539284706116,
-0.008930564858019352,
0.1574520617723465,
-0.09063117951154709,
-0.1197347566485405,
0.0484822541475296,
-0.0630258172750473,
0.14611664414405823,
0.02818998321890831,
-0.034195899963378906,
0.08192719519138336,
-0.16197480261325836,
-0.0339224636554718,
-0.028060175478458405,
-0.00723256217315793,
0.10402708500623703,
0.09210056811571121,
-0.06505972892045975,
0.0468456894159317,
0.018044332042336464,
-0.04152099788188934,
-0.038068629801273346,
-0.052533939480781555,
-0.11844942718744278,
0.0012227081460878253,
-0.08217319846153259,
0.03508223965764046,
-0.013970605097711086,
-0.06411232799291611,
-0.013696790672838688,
-0.1674257218837738,
-0.004452073480933905,
0.09094837307929993,
0.011956306174397469,
-0.022260021418333054,
-0.09328074008226395,
0.023445552214980125,
-0.010471045039594173,
-0.03440795838832855,
-0.14324605464935303,
-0.03640478104352951,
0.012563892640173435,
-0.13831567764282227,
0.016943639144301414,
-0.11671459674835205,
0.0554199293255806,
0.020001012831926346,
-0.06914778798818588,
-0.01669548638164997,
-0.022785015404224396,
0.021327996626496315,
-0.04889942333102226,
-0.24583099782466888,
-0.017665192484855652,
-0.04293544590473175,
0.1494484841823578,
-0.22400592267513275,
0.04160754382610321,
0.056638024747371674,
0.12506844103336334,
-0.012923712842166424,
-0.05817665532231331,
0.023863838985562325,
-0.07131504267454147,
-0.024320119991898537,
-0.05982694774866104,
-0.013311131857335567,
-0.015601590275764465,
-0.05537383258342743,
0.022339439019560814,
-0.11295162886381149,
-0.04536622762680054,
0.10713870078325272,
0.07675379514694214,
-0.16534005105495453,
-0.03820289671421051,
-0.03140741214156151,
-0.08164740353822708,
-0.09000084549188614,
-0.06165561079978943,
0.10615158081054688,
0.05043869838118553,
0.03370879963040352,
-0.08126258850097656,
-0.08329309523105621,
0.008826100267469883,
-0.024882493540644646,
-0.02908565104007721,
0.10402064770460129,
0.059376977384090424,
-0.11822085827589035,
0.09933172166347504,
0.07895755022764206,
0.0057035754434764385,
0.10207298398017883,
-0.020556187257170677,
-0.11644016206264496,
-0.046322669833898544,
0.041579440236091614,
0.0066756634041666985,
0.17341621220111847,
-0.07063884288072586,
0.06644845753908157,
0.0392342172563076,
-0.021938681602478027,
0.05796538665890694,
-0.0984586775302887,
0.013013722375035286,
-0.00014002474199514836,
-0.015127699822187424,
0.0012232492445036769,
-0.030704092234373093,
0.02451329305768013,
0.07556341588497162,
0.042123571038246155,
0.036042533814907074,
0.043708011507987976,
-0.038218654692173004,
-0.12722530961036682,
0.1888585090637207,
-0.11132288724184036,
-0.2106482982635498,
-0.1631026268005371,
0.0538061261177063,
0.03626300022006035,
-0.026809824630618095,
0.008883481845259666,
-0.03901756927371025,
-0.09378311783075333,
-0.08165492862462997,
-0.003493528114631772,
0.03679237142205238,
-0.0676792711019516,
-0.07970349490642548,
0.06169610098004341,
0.05629400163888931,
-0.12466651201248169,
0.04076872020959854,
0.055601224303245544,
-0.030898114666342735,
0.00625052023679018,
0.07371595501899719,
0.07530564069747925,
0.14935843646526337,
-0.014906415715813637,
-0.00743898143991828,
0.052912477403879166,
0.2596709728240967,
-0.15353995561599731,
0.09615891426801682,
0.11230926960706711,
-0.06936117261648178,
0.07784450799226761,
0.179547980427742,
0.032167382538318634,
-0.10329442471265793,
0.03954490274190903,
0.03297366574406624,
-0.018733708187937737,
-0.2761574685573578,
-0.05350080132484436,
-0.005191644188016653,
-0.10629693418741226,
0.06754535436630249,
0.07840600609779358,
0.08623511344194412,
0.040893230587244034,
-0.06452789157629013,
-0.09204737842082977,
0.029733341187238693,
0.08125974237918854,
-0.02590303122997284,
0.004223145544528961,
0.07660869508981705,
-0.01868366077542305,
0.01718960888683796,
0.10475799441337585,
-0.003060521325096488,
0.18711476027965546,
0.027790617197752,
0.09382420778274536,
0.09295517206192017,
0.10561932623386383,
-0.012411389499902725,
0.020525166764855385,
0.019523248076438904,
0.020536812022328377,
0.008625704795122147,
-0.08736921101808548,
0.03408323973417282,
0.11388731002807617,
0.04893956705927849,
0.0401281900703907,
0.011612595058977604,
-0.04936189576983452,
0.05163692310452461,
0.17860716581344604,
-0.000170899074873887,
-0.20349018275737762,
-0.0693155974149704,
0.058504924178123474,
-0.0795617401599884,
-0.13149207830429077,
-0.018380293622612953,
0.0517018586397171,
-0.1681850403547287,
0.021011916920542717,
-0.04702027514576912,
0.09291640669107437,
-0.07808839529752731,
-0.04085437208414078,
0.08330607414245605,
0.07489974051713943,
-0.018271667882800102,
0.07440989464521408,
-0.1822451949119568,
0.130047008395195,
0.024025702849030495,
0.07504849880933762,
-0.08977258950471878,
0.10318286716938019,
0.019651198759675026,
-0.016404079273343086,
0.15247845649719238,
0.00700495345517993,
-0.035746168345212936,
-0.061536550521850586,
-0.11217644065618515,
-0.010082807391881943,
0.09073230624198914,
-0.11768040060997009,
0.06739463657140732,
-0.008377942256629467,
-0.021237369626760483,
0.010276824235916138,
-0.06931021064519882,
-0.1321897804737091,
-0.16883379220962524,
0.057229045778512955,
-0.12265047430992126,
0.0442219115793705,
-0.09883608669042587,
-0.0708678811788559,
-0.008518463931977749,
0.16887374222278595,
-0.192191019654274,
-0.07580472528934479,
-0.14189769327640533,
-0.08355502784252167,
0.17208366096019745,
-0.04507419094443321,
0.07428380846977234,
0.022840114310383797,
0.1582527905702591,
0.027413588017225266,
0.001366468844935298,
0.09746744483709335,
-0.08509296178817749,
-0.18868382275104523,
-0.05685028061270714,
0.15007854998111725,
0.16047607362270355,
0.03972518816590309,
-0.01171641331166029,
0.00759653327986598,
-0.053331293165683746,
-0.12062567472457886,
0.011744651943445206,
0.16191375255584717,
0.10482793301343918,
0.006793193053454161,
-0.02706596627831459,
-0.11145397275686264,
-0.06330008804798126,
-0.06478084623813629,
-0.002286971779540181,
0.19048772752285004,
-0.06574860960245132,
0.15543584525585175,
0.1109643280506134,
-0.05604147911071777,
-0.20484483242034912,
0.04608650133013725,
0.06657455116510391,
0.023484762758016586,
0.06219375506043434,
-0.16893428564071655,
0.10238359868526459,
0.024783534929156303,
-0.06445116549730301,
0.14760766923427582,
-0.142522394657135,
-0.14935795962810516,
0.0961783304810524,
0.05048270896077156,
-0.23321887850761414,
-0.11207089573144913,
-0.09536226838827133,
-0.028544528409838676,
-0.11171527951955795,
0.0760783776640892,
-0.024607311934232712,
0.013505135662853718,
0.030618060380220413,
0.034116797149181366,
0.019545556977391243,
-0.046969059854745865,
0.2058962881565094,
-0.011812075041234493,
0.031025355681777,
-0.04866284877061844,
-0.08003591001033783,
0.03948293626308441,
-0.05390021950006485,
0.09585001319646835,
-0.013421738520264626,
0.0265596155077219,
-0.12655775249004364,
-0.04175291210412979,
-0.06944926083087921,
0.031202754005789757,
-0.10094862431287766,
-0.08246374875307083,
-0.048856448382139206,
0.10299462825059891,
0.08941899985074997,
-0.04341145604848862,
0.00008339050691574812,
-0.07415939122438431,
0.04515631124377251,
0.21254324913024902,
0.1935090869665146,
0.05401868000626564,
-0.06539389491081238,
0.011519267223775387,
-0.021569542586803436,
0.04727421700954437,
-0.2247365415096283,
0.05358831211924553,
0.04411378502845764,
0.022826576605439186,
0.09642569720745087,
-0.022250307723879814,
-0.15034323930740356,
-0.0656452402472496,
0.07169666141271591,
-0.04211362823843956,
-0.15292060375213623,
-0.03115718625485897,
0.02961842529475689,
-0.20632213354110718,
-0.03738311678171158,
0.022530773654580116,
-0.018216492608189583,
-0.043114449828863144,
0.016917221248149872,
0.08529895544052124,
-0.02087024226784706,
0.13345685601234436,
0.08957752585411072,
0.09216758608818054,
-0.10169534385204315,
0.07263094931840897,
0.06009650230407715,
-0.050556477159261703,
0.031720779836177826,
0.09272869676351547,
-0.04466380923986435,
-0.03710941597819328,
0.09317772835493088,
0.06559107452630997,
0.035064175724983215,
-0.04930846765637398,
-0.0033775418996810913,
-0.04918081313371658,
0.05591424182057381,
0.11969546973705292,
0.04869379103183746,
0.003908928018063307,
0.05273585394024849,
0.026378637179732323,
-0.0883166491985321,
0.1158120334148407,
0.05843094363808632,
0.02394208312034607,
-0.049317795783281326,
-0.026299387216567993,
-0.0013593442272394896,
-0.006479601375758648,
-0.01868307776749134,
-0.008035797625780106,
-0.08306128531694412,
-0.01171721238642931,
-0.12620829045772552,
0.04390465095639229,
-0.09188356250524521,
0.013429909013211727,
0.022517696022987366,
-0.049229398369789124,
0.0006916942656971514,
0.0176959540694952,
-0.07154778391122818,
-0.05301358178257942,
-0.008699357509613037,
0.10766588896512985,
-0.12344101816415787,
0.03539144992828369,
0.0804985836148262,
-0.10469279438257217,
0.07710271328687668,
0.007031033281236887,
0.005612345412373543,
0.022744860500097275,
-0.16864150762557983,
0.05949518084526062,
-0.02903292141854763,
-0.007960217073559761,
0.024716878309845924,
-0.23548151552677155,
-0.011425469070672989,
-0.03641229122877121,
-0.03063340112566948,
0.014378620311617851,
-0.02933029644191265,
-0.12826940417289734,
0.07657664269208908,
-0.009620917029678822,
-0.07099192589521408,
-0.025762317702174187,
0.035869766026735306,
0.11137785017490387,
-0.027072051540017128,
0.1433473378419876,
-0.020530300214886665,
0.06747234612703323,
-0.17002108693122864,
-0.0064101857133209705,
-0.016799164935946465,
0.042844340205192566,
-0.017992954701185226,
-0.015704413875937462,
0.058607108891010284,
-0.030620155856013298,
0.19530117511749268,
-0.03406386077404022,
0.05592560023069382,
0.05420104041695595,
0.006897786166518927,
-0.005235535558313131,
0.08602031320333481,
0.0674787238240242,
-0.018315263092517853,
0.012456084601581097,
0.041739195585250854,
-0.009200173430144787,
-0.04386542737483978,
-0.1520599126815796,
0.06064756214618683,
0.16119173169136047,
0.03741957247257233,
0.01846368983387947,
0.0556517019867897,
-0.1081010103225708,
-0.07354220002889633,
0.134055957198143,
-0.0037154064048081636,
-0.03955170512199402,
-0.07482743263244629,
0.1456473171710968,
0.12226816266775131,
-0.2028440684080124,
0.08005663752555847,
-0.06430520117282867,
-0.07309237122535706,
-0.1145973950624466,
-0.15538856387138367,
-0.06110803410410881,
-0.051249533891677856,
-0.01601872406899929,
-0.0641443133354187,
0.05452084168791771,
0.08872498571872711,
0.006302941124886274,
-0.021787574514746666,
0.10528529435396194,
0.0038734038826078176,
-0.02762324921786785,
0.034976497292518616,
0.06584593653678894,
0.01988472044467926,
-0.09605155885219574,
0.011486846022307873,
-0.0022990719880908728,
0.022245321422815323,
0.06278849393129349,
0.0024476663675159216,
-0.03377585485577583,
-0.0040506636723876,
-0.029394136741757393,
-0.11093170195817947,
0.044652730226516724,
-0.019515272229909897,
-0.028262600302696228,
0.13903376460075378,
0.02630247361958027,
-0.003486504778265953,
-0.023687375709414482,
0.23213301599025726,
-0.08277353644371033,
-0.08822204172611237,
-0.16660727560520172,
0.06670195609331131,
-0.05118541046977043,
0.026055844500660896,
0.04193805530667305,
-0.11514432728290558,
0.02673569694161415,
0.13941872119903564,
0.13982562720775604,
-0.012767483480274677,
0.011560298502445221,
0.042639147490262985,
-0.003045852994546294,
-0.04272124916315079,
0.017718058079481125,
0.048053305596113205,
0.11748947948217392,
-0.055739015340805054,
0.08068620413541794,
-0.007051908876746893,
-0.07979767769575119,
-0.004068160895258188,
0.10759461671113968,
-0.0066666859202086926,
0.009513513185083866,
-0.06542395800352097,
0.1420755386352539,
-0.061293136328458786,
-0.23070363700389862,
0.05206836014986038,
-0.0649845078587532,
-0.16111242771148682,
-0.03745974600315094,
0.018520919606089592,
-0.022760067135095596,
0.013370173051953316,
0.07987762242555618,
-0.040034808218479156,
0.18289142847061157,
0.04053264483809471,
-0.06730475276708603,
-0.07293884456157684,
0.0719963014125824,
-0.12805137038230896,
0.28363844752311707,
0.01739502139389515,
0.06584255397319794,
0.10554805397987366,
-0.017689306288957596,
-0.13163237273693085,
0.025880809873342514,
0.0994238406419754,
-0.07768155634403229,
0.0731414407491684,
0.19108183681964874,
-0.001895581604912877,
0.13861946761608124,
0.05940021201968193,
-0.040656447410583496,
0.026879940181970596,
-0.11860781908035278,
-0.06058025360107422,
-0.11138372868299484,
0.0808485597372055,
-0.07967403531074524,
0.16466905176639557,
0.1387745887041092,
-0.07049159705638885,
0.00014616368571296334,
-0.022739756852388382,
0.0835975706577301,
-0.006697655655443668,
0.10973099619150162,
0.007436915300786495,
-0.20406930148601532,
0.0314500518143177,
0.022381901741027832,
0.10284528881311417,
-0.21157197654247284,
-0.06498019397258759,
0.06292314827442169,
-0.030605090782046318,
-0.056584782898426056,
0.11501917988061905,
0.04757354035973549,
0.03882963955402374,
-0.03858773410320282,
-0.03975297138094902,
-0.02144400030374527,
0.13411201536655426,
-0.11429842561483383,
-0.015013635158538818
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-cased-lora-592K-snli-model3
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8978
- Accuracy: 0.645
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 256
- eval_batch_size: 256
- seed: 46
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.6372 | 1.0 | 2146 | 0.5562 | 0.7846 |
| 0.5816 | 2.0 | 4292 | 0.5116 | 0.8021 |
| 0.564 | 3.0 | 6438 | 0.4957 | 0.8072 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-base-cased-lora-592K-snli-model3", "results": []}]} | text-classification | varun-v-rao/bert-base-cased-lora-592K-snli-model3 | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:bert-base-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T05:23:02+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-base-cased-lora-592K-snli-model3
=====================================
This model is a fine-tuned version of bert-base-cased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8978
* Accuracy: 0.645
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 256
* eval\_batch\_size: 256
* seed: 46
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 46\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 46\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
67,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 46\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.08975468575954437,
0.10616015642881393,
-0.002115728100761771,
0.10651184618473053,
0.15027600526809692,
0.02154955267906189,
0.15638510882854462,
0.10905218869447708,
-0.07903268933296204,
0.046616457402706146,
0.12526820600032806,
0.12785860896110535,
0.01492181234061718,
0.12948037683963776,
-0.062362026423215866,
-0.22526569664478302,
0.008811607956886292,
0.03307228162884712,
-0.07077199965715408,
0.1171075850725174,
0.09044181555509567,
-0.13040515780448914,
0.0908985435962677,
-0.009510611183941364,
-0.1800641119480133,
0.01934257335960865,
0.02163836359977722,
-0.0518675372004509,
0.13552074134349823,
0.03550232574343681,
0.13932137191295624,
0.02100575901567936,
0.09128114581108093,
-0.20905627310276031,
0.011132368817925453,
0.06402728706598282,
-0.006639648694545031,
0.0835968479514122,
0.03438691794872284,
0.009953980334103107,
0.0885772779583931,
-0.08148182928562164,
0.06273095309734344,
0.018205199390649796,
-0.11832299828529358,
-0.1931673139333725,
-0.07279003411531448,
0.04209701716899872,
0.0931781753897667,
0.07784795016050339,
-0.011683867312967777,
0.12427674233913422,
-0.05634889379143715,
0.08932770043611526,
0.20802877843379974,
-0.32232293486595154,
-0.06575185060501099,
0.04952911660075188,
0.026206372305750847,
0.08543485403060913,
-0.10460663586854935,
-0.0175317469984293,
0.06899445503950119,
0.02798686921596527,
0.1314900666475296,
-0.027933314442634583,
-0.07309652119874954,
0.007951313629746437,
-0.15035033226013184,
-0.01650584302842617,
0.14662523567676544,
0.051274243742227554,
-0.04181741550564766,
-0.04396244138479233,
-0.06374932080507278,
-0.15204128623008728,
-0.040663719177246094,
-0.028014419600367546,
0.052061136811971664,
-0.020455939695239067,
-0.05876882001757622,
-0.009968011640012264,
-0.10820241272449493,
-0.08046823740005493,
-0.05767718330025673,
0.1338314712047577,
0.04043033719062805,
0.0016475229058414698,
-0.012993419542908669,
0.09743960946798325,
-0.034783992916345596,
-0.1195751428604126,
0.0241581741720438,
0.023712830618023872,
0.010890903882682323,
-0.057420697063207626,
-0.053415555506944656,
-0.06481385231018066,
0.027619868516921997,
0.13040903210639954,
-0.05675308778882027,
0.05359657481312752,
0.009073197841644287,
0.045944176614284515,
-0.10295704007148743,
0.1581585705280304,
-0.04044671729207039,
-0.024818725883960724,
0.016744231805205345,
0.0801180750131607,
0.041206393390893936,
-0.0006396942771971226,
-0.127021923661232,
0.02872992493212223,
0.11060633510351181,
0.01609026826918125,
-0.08132573962211609,
0.08058499544858932,
-0.04993409663438797,
0.0023694343399256468,
0.01743166893720627,
-0.08693704754114151,
0.02962275967001915,
0.00729961134493351,
-0.052109044045209885,
-0.07536900043487549,
0.029720254242420197,
0.018130537122488022,
0.0025338244158774614,
0.10330872237682343,
-0.08766338974237442,
0.008830524049699306,
-0.09399528801441193,
-0.1231406033039093,
0.020137449726462364,
-0.07591976225376129,
0.021944276988506317,
-0.11079021543264389,
-0.15501077473163605,
-0.007711485959589481,
0.05680935084819794,
-0.02414562739431858,
-0.03010040894150734,
-0.05615615472197533,
-0.07287909090518951,
0.012297015637159348,
-0.017674043774604797,
0.09506475180387497,
-0.06559181958436966,
0.09347651153802872,
0.037400662899017334,
0.06546207517385483,
-0.04845843464136124,
0.03824036568403244,
-0.09933142364025116,
0.02632889151573181,
-0.18146522343158722,
0.012754866853356361,
-0.0689721331000328,
0.05879644677042961,
-0.08297780901193619,
-0.07572925835847855,
-0.01113432738929987,
0.015143808908760548,
0.07303614914417267,
0.09152316302061081,
-0.16294777393341064,
-0.06329727917909622,
0.17575445771217346,
-0.09640701860189438,
-0.13961701095104218,
0.12358826398849487,
-0.05965524539351463,
0.05100912228226662,
0.06582751125097275,
0.18161526322364807,
0.06615894287824631,
-0.09727486222982407,
0.0024353372864425182,
0.0022811032831668854,
0.06078306958079338,
-0.05217744782567024,
0.061405230313539505,
0.0019511448917910457,
-0.015635982155799866,
0.016550064086914062,
-0.04773733764886856,
0.05028790980577469,
-0.08052624762058258,
-0.08466234058141708,
-0.03761202096939087,
-0.10256537050008774,
0.045572586357593536,
0.05734262615442276,
0.06360302120447159,
-0.1167069599032402,
-0.08945576101541519,
0.07129719853401184,
0.07295159250497818,
-0.07275514304637909,
0.017318667843937874,
-0.0724935233592987,
0.07806701213121414,
-0.057589638978242874,
-0.017877502366900444,
-0.1509973108768463,
-0.048147886991500854,
0.01571071706712246,
-0.0005102212307974696,
0.013058323413133621,
0.00795143935829401,
0.0716741532087326,
0.07741668820381165,
-0.06794875115156174,
-0.025597047060728073,
-0.013634279370307922,
0.018457695841789246,
-0.13101635873317719,
-0.20681072771549225,
-0.015783270820975304,
-0.03218279778957367,
0.135115846991539,
-0.24306407570838928,
0.05193314328789711,
0.000461672549135983,
0.07871612161397934,
0.034693557769060135,
-0.0038182372227311134,
-0.04537709057331085,
0.07632122188806534,
-0.04184575751423836,
-0.05389752611517906,
0.06134195253252983,
0.012484404258430004,
-0.08099044859409332,
-0.03867413476109505,
-0.145868182182312,
0.19185428321361542,
0.13405248522758484,
-0.0964093878865242,
-0.08514855057001114,
-0.013992950320243835,
-0.0418010838329792,
-0.02949708327651024,
-0.04751954227685928,
0.005158267915248871,
0.1349652111530304,
-0.018147356808185577,
0.15092860162258148,
-0.07989569008350372,
-0.031565919518470764,
0.02462262287735939,
-0.04740521311759949,
0.009071261622011662,
0.11215110868215561,
0.11885253340005875,
-0.11148232966661453,
0.15519261360168457,
0.17364640533924103,
-0.09847839176654816,
0.13661012053489685,
-0.04197793826460838,
-0.057142049074172974,
-0.02080605924129486,
0.007970627397298813,
0.00445944769307971,
0.11915834993124008,
-0.13209126889705658,
-0.0033189449459314346,
0.0027920319698750973,
0.014156402088701725,
0.018960295245051384,
-0.22667475044727325,
-0.02762700617313385,
0.03744736313819885,
-0.04968569427728653,
0.011133220046758652,
-0.031640056520700455,
-0.011416145600378513,
0.10086461901664734,
-0.0018590387189760804,
-0.08772894740104675,
0.0427144356071949,
-0.004870983771979809,
-0.09067622572183609,
0.2125266194343567,
-0.08306378126144409,
-0.13451316952705383,
-0.13278728723526,
-0.06435099989175797,
-0.03399175405502319,
0.02468900755047798,
0.06224432587623596,
-0.08108369261026382,
-0.03894798457622528,
-0.1078135073184967,
-0.0005481704138219357,
0.033398475497961044,
0.02916608564555645,
0.01028137095272541,
0.004358408506959677,
0.08345216512680054,
-0.10189889371395111,
-0.011961033567786217,
-0.05281640961766243,
-0.06068000942468643,
0.03005433827638626,
0.026720484718680382,
0.11528736352920532,
0.1479106843471527,
-0.03288784250617027,
-0.0027727338019758463,
-0.03455507382750511,
0.21800114214420319,
-0.05354411527514458,
-0.018683746457099915,
0.11824321746826172,
-0.02726800926029682,
0.04400326684117317,
0.1371576189994812,
0.0643906369805336,
-0.09400641918182373,
0.024267785251140594,
0.04440254718065262,
-0.027842877432703972,
-0.21419452130794525,
-0.035842131823301315,
-0.034372735768556595,
-0.0004977716598659754,
0.09731191396713257,
0.035370491445064545,
0.02294257842004299,
0.06868989020586014,
0.035636529326438904,
0.07462761551141739,
-0.015616917051374912,
0.0648229718208313,
0.11492762714624405,
0.039442289620637894,
0.12586286664009094,
-0.05241202190518379,
-0.06183011084794998,
0.03287846967577934,
0.0031059079337865114,
0.20262837409973145,
0.028331264853477478,
0.1262318640947342,
0.06192528456449509,
0.16149437427520752,
0.0031268077436834574,
0.06909922510385513,
-0.012382621876895428,
-0.04910145699977875,
-0.010219665244221687,
-0.05060940980911255,
-0.031777121126651764,
0.039112672209739685,
-0.09632597863674164,
0.05940507724881172,
-0.11504489928483963,
0.01797722838819027,
0.05909613519906998,
0.23868702352046967,
0.045325007289648056,
-0.3208633065223694,
-0.08667007833719254,
0.02283337153494358,
-0.027141664177179337,
-0.022685319185256958,
0.03588205575942993,
0.13618560135364532,
-0.04938305541872978,
0.023560699075460434,
-0.07039477676153183,
0.07826432585716248,
-0.04298817366361618,
0.04512185975909233,
0.06636360287666321,
0.08666259795427322,
-0.009956269524991512,
0.07107082009315491,
-0.2817591726779938,
0.2754099369049072,
0.013225751928985119,
0.07513601332902908,
-0.0517200343310833,
0.0017645481275394559,
0.032595306634902954,
0.09074157476425171,
0.07443397492170334,
-0.022736337035894394,
-0.07280571013689041,
-0.20087800920009613,
-0.05253848060965538,
0.03205551952123642,
0.09358316659927368,
-0.028814123943448067,
0.09706577658653259,
-0.034513238817453384,
0.003154069185256958,
0.09169626981019974,
-0.014597700908780098,
-0.08465685695409775,
-0.09736926853656769,
-0.026046661660075188,
0.04005993902683258,
-0.03263162076473236,
-0.08340363204479218,
-0.10497772693634033,
-0.13394136726856232,
0.1549481749534607,
-0.05821920558810234,
-0.021685132756829262,
-0.09267275035381317,
0.06970733404159546,
0.039883244782686234,
-0.07577261328697205,
0.049753572791814804,
0.009995365515351295,
0.08634842187166214,
0.02456853725016117,
-0.05957129970192909,
0.12947793304920197,
-0.07567039132118225,
-0.16400709748268127,
-0.07779832184314728,
0.0989760309457779,
0.026521041989326477,
0.04551001265645027,
-0.00044320651795715094,
0.001911877654492855,
-0.006728088948875666,
-0.07712899148464203,
0.023690469563007355,
-0.008504231460392475,
0.0673864483833313,
0.006620478816330433,
-0.08646153658628464,
-0.006260995287448168,
-0.05266096442937851,
-0.02979486994445324,
0.1659231334924698,
0.26737290620803833,
-0.09206171333789825,
-0.002858899999409914,
0.06684326380491257,
-0.07458163052797318,
-0.2145882099866867,
0.04764305800199509,
0.02964530698955059,
0.003564254380762577,
0.04327533394098282,
-0.15396931767463684,
0.1221674457192421,
0.10234858840703964,
-0.0277739055454731,
0.10473977029323578,
-0.2843092978000641,
-0.13102936744689941,
0.13863584399223328,
0.16695822775363922,
0.1102331131696701,
-0.15910403430461884,
-0.03046971932053566,
-0.03132506087422371,
-0.10366912186145782,
0.11330517381429672,
-0.12572644650936127,
0.105059415102005,
-0.004228614736348391,
0.05215349420905113,
0.0009341603727079928,
-0.05740547552704811,
0.11806754767894745,
-0.018280858173966408,
0.11501899361610413,
-0.061423368752002716,
-0.023678716272115707,
0.04496937617659569,
-0.05281033739447594,
0.018947700038552284,
-0.11728514730930328,
0.030052339658141136,
-0.0513225719332695,
-0.03125876188278198,
-0.04687866196036339,
0.03565540537238121,
-0.039902009069919586,
-0.06531701982021332,
-0.03631959110498428,
0.027898982167243958,
0.03591699153184891,
-0.012399809435009956,
0.1355956792831421,
0.009132024832069874,
0.15731102228164673,
0.14063774049282074,
0.08402823656797409,
-0.06207430735230446,
-0.01700216718018055,
0.0031562147196382284,
-0.03768053650856018,
0.06791681051254272,
-0.14877977967262268,
0.039164185523986816,
0.12259180098772049,
0.003413819707930088,
0.15118853747844696,
0.07615535706281662,
-0.031245091930031776,
0.005258797202259302,
0.07231192290782928,
-0.16409847140312195,
-0.10718996077775955,
-0.00011903208360308781,
-0.03271365538239479,
-0.11612814664840698,
0.0704171359539032,
0.11380022764205933,
-0.07663393765687943,
0.00643904646858573,
-0.00811956636607647,
0.012204945087432861,
-0.0442478284239769,
0.17925144731998444,
0.07166757434606552,
0.0483490452170372,
-0.07260409742593765,
0.07988159358501434,
0.041600171476602554,
-0.07598230242729187,
0.02072090655565262,
0.038701754063367844,
-0.08550659567117691,
-0.04959774389863014,
0.06339515000581741,
0.2032902091741562,
-0.014309637248516083,
-0.06569457799196243,
-0.14277727901935577,
-0.11977273970842361,
0.055084168910980225,
0.195176362991333,
0.1001591682434082,
0.007706294301897287,
-0.034205805510282516,
0.022763442248106003,
-0.11897775530815125,
0.11686467379331589,
0.022982046008110046,
0.08273258060216904,
-0.15081772208213806,
0.12540005147457123,
0.002153492532670498,
0.0032751690596342087,
-0.02670520357787609,
0.050064485520124435,
-0.12426622956991196,
-0.004988601431250572,
-0.12790633738040924,
-0.008436041884124279,
-0.02376672998070717,
0.007487154565751553,
0.010055136866867542,
-0.049240242689847946,
-0.06754013150930405,
0.013777622953057289,
-0.1026679202914238,
-0.012755390256643295,
0.038533929735422134,
0.06327050924301147,
-0.12453850358724594,
-0.03487658128142357,
0.022552674636244774,
-0.06348856538534164,
0.062371306121349335,
0.014390935190021992,
0.02489006519317627,
0.05815086141228676,
-0.17849135398864746,
0.0393817238509655,
0.07318886369466782,
0.011316148564219475,
0.04383915662765503,
-0.09084651619195938,
-0.009061305783689022,
-0.0005053867353126407,
0.046906061470508575,
0.0182412788271904,
0.09372110664844513,
-0.126869335770607,
0.009030509740114212,
-0.029114771634340286,
-0.06571518629789352,
-0.04901431128382683,
0.02386549487709999,
0.08820577710866928,
-0.004025962669402361,
0.20181511342525482,
-0.10132776200771332,
0.006976678501814604,
-0.20314230024814606,
0.00789363868534565,
0.0012930475641041994,
-0.106998510658741,
-0.12361384183168411,
-0.05737832188606262,
0.04768342524766922,
-0.06122910603880882,
0.15231890976428986,
0.02334262616932392,
0.016890201717615128,
0.03549060970544815,
-0.051069144159555435,
0.03832901641726494,
0.03585295006632805,
0.22745992243289948,
0.02455746941268444,
-0.04394255951046944,
0.017560653388500214,
0.035789649933576584,
0.1127927154302597,
0.08060242980718613,
0.1731390655040741,
0.16990558803081512,
-0.0505884625017643,
0.10481687635183334,
0.049214109778404236,
-0.05576717108488083,
-0.1309090405702591,
0.06639667600393295,
-0.052455026656389236,
0.09880893677473068,
-0.0272000003606081,
0.2221895158290863,
0.09078975021839142,
-0.15395605564117432,
0.010950906202197075,
-0.0632527694106102,
-0.08103708177804947,
-0.1179790124297142,
-0.05453107878565788,
-0.0980776697397232,
-0.14981582760810852,
-0.0018259070347994566,
-0.11463113129138947,
-0.0016302186995744705,
0.11025433242321014,
0.0052413661032915115,
-0.01589798368513584,
0.17653296887874603,
0.009706978686153889,
0.04079929366707802,
0.05399748310446739,
0.004738802555948496,
-0.035063911229372025,
-0.10778321325778961,
-0.08787930011749268,
0.002974873874336481,
-0.02460838109254837,
0.02019956335425377,
-0.052288442850112915,
-0.025253504514694214,
0.042627573013305664,
-0.005758466199040413,
-0.09446325153112411,
0.010333675891160965,
0.02142992429435253,
0.05000695213675499,
0.044655513018369675,
0.009010453708469868,
0.006014971062541008,
0.003691558726131916,
0.2075524479150772,
-0.07588110864162445,
-0.060738541185855865,
-0.10120821744203568,
0.2257085144519806,
0.028104573488235474,
0.026510465890169144,
0.011193926446139812,
-0.08553516864776611,
0.029762497171759605,
0.23015081882476807,
0.19017483294010162,
-0.08271694928407669,
0.0018624432850629091,
-0.003015509806573391,
-0.01152925193309784,
-0.04063435643911362,
0.10493901371955872,
0.12086714059114456,
0.015784932300448418,
-0.07610343396663666,
-0.051656775176525116,
-0.03512183204293251,
-0.0010697406250983477,
-0.039643287658691406,
0.054742589592933655,
0.04082470014691353,
0.015771610662341118,
-0.05100015178322792,
0.049527522176504135,
-0.025842269882559776,
-0.11624496430158615,
0.059034574776887894,
-0.1946200579404831,
-0.14790064096450806,
-0.00911834929138422,
0.12010776251554489,
-0.018765773624181747,
0.04946129769086838,
-0.03460750728845596,
-0.002132222754880786,
0.07263925671577454,
-0.02526695653796196,
-0.06901461631059647,
-0.07960879802703857,
0.05743652582168579,
-0.0986751839518547,
0.23881743848323822,
-0.03687530383467674,
0.04607686772942543,
0.1328520029783249,
0.04388044774532318,
-0.0704837217926979,
0.08721467852592468,
0.045209404081106186,
-0.07716314494609833,
0.02466019056737423,
0.061772726476192474,
-0.043165504932403564,
0.12371892482042313,
0.054942794144153595,
-0.13265852630138397,
0.022026872262358665,
-0.07867640256881714,
-0.0902509093284607,
-0.05600441247224808,
-0.03273152559995651,
-0.06549742072820663,
0.1334046870470047,
0.18779461085796356,
-0.028869712725281715,
0.010656151920557022,
-0.048686906695365906,
0.024312816560268402,
0.0642554759979248,
0.04215973988175392,
-0.030818665400147438,
-0.23482738435268402,
0.03187551721930504,
0.07881201058626175,
-0.010018162429332733,
-0.2904622554779053,
-0.0872965082526207,
-0.00121600111015141,
-0.045279886573553085,
-0.10109449923038483,
0.07942860573530197,
0.1279689520597458,
0.06111660599708557,
-0.06875674426555634,
-0.11043508350849152,
-0.07366673648357391,
0.15712495148181915,
-0.12689046561717987,
-0.10282494127750397
] |
null | null | transformers | # MarcoroCapy-7B
This model is a DPO fine tune of [mlabonne/Marcoro14-7B-slerp](https://huggingface.co/mlabonne/Marcoro14-7B-slerp) on [argilla/distilabel-capybara-dpo-7k-binarized](https://huggingface.co/datasets/argilla/distilabel-capybara-dpo-7k-binarized)
<div align="center">

[<img src="https://raw.githubusercontent.com/argilla-io/distilabel/main/docs/assets/distilabel-badge-dark.png" alt="Built with Distilabel" width="200" height="32"/>](https://github.com/argilla-io/distilabel)
</div>
## Process
+ Realigned the chat template to ChatML
+ Completed 1 Epoch
+ 5e-5 learning rate
+ Training time was about 4.5 hours on 1 H100
+ Cost was ~$20
## GGUF
TODO
## Evaluations
TODO
| {"library_name": "transformers", "tags": []} | text-generation | macadeliccc/MarcoroCapy-7B | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T05:26:34+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MarcoroCapy-7B
This model is a DPO fine tune of mlabonne/Marcoro14-7B-slerp on argilla/distilabel-capybara-dpo-7k-binarized
<div align="center">
!image/webp
<img src="URL alt="Built with Distilabel" width="200" height="32"/>
</div>
## Process
+ Realigned the chat template to ChatML
+ Completed 1 Epoch
+ 5e-5 learning rate
+ Training time was about 4.5 hours on 1 H100
+ Cost was ~$20
## GGUF
TODO
## Evaluations
TODO
| [
"# MarcoroCapy-7B\n\nThis model is a DPO fine tune of mlabonne/Marcoro14-7B-slerp on argilla/distilabel-capybara-dpo-7k-binarized\n\n<div align=\"center\"> \n\n!image/webp\n\n<img src=\"URL alt=\"Built with Distilabel\" width=\"200\" height=\"32\"/>\n\n</div>",
"## Process\n\n+ Realigned the chat template to ChatML \n+ Completed 1 Epoch\n+ 5e-5 learning rate\n+ Training time was about 4.5 hours on 1 H100\n+ Cost was ~$20",
"## GGUF\n\nTODO",
"## Evaluations\n\nTODO"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MarcoroCapy-7B\n\nThis model is a DPO fine tune of mlabonne/Marcoro14-7B-slerp on argilla/distilabel-capybara-dpo-7k-binarized\n\n<div align=\"center\"> \n\n!image/webp\n\n<img src=\"URL alt=\"Built with Distilabel\" width=\"200\" height=\"32\"/>\n\n</div>",
"## Process\n\n+ Realigned the chat template to ChatML \n+ Completed 1 Epoch\n+ 5e-5 learning rate\n+ Training time was about 4.5 hours on 1 H100\n+ Cost was ~$20",
"## GGUF\n\nTODO",
"## Evaluations\n\nTODO"
] | [
47,
91,
41,
6,
5
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MarcoroCapy-7B\n\nThis model is a DPO fine tune of mlabonne/Marcoro14-7B-slerp on argilla/distilabel-capybara-dpo-7k-binarized\n\n<div align=\"center\"> \n\n!image/webp\n\n<img src=\"URL alt=\"Built with Distilabel\" width=\"200\" height=\"32\"/>\n\n</div>## Process\n\n+ Realigned the chat template to ChatML \n+ Completed 1 Epoch\n+ 5e-5 learning rate\n+ Training time was about 4.5 hours on 1 H100\n+ Cost was ~$20## GGUF\n\nTODO## Evaluations\n\nTODO"
] | [
-0.14279377460479736,
0.010144504718482494,
-0.0007890493143349886,
0.020312225446105003,
0.12760287523269653,
-0.04121774062514305,
0.14451441168785095,
0.05564195290207863,
-0.016270365566015244,
0.0357138067483902,
0.06444784998893738,
0.10640467703342438,
0.05382278189063072,
0.20492969453334808,
-0.11337487399578094,
-0.22819288074970245,
0.06786365061998367,
0.007044428028166294,
0.015516088344156742,
0.0931447297334671,
0.12585166096687317,
-0.07392016798257828,
0.06829650700092316,
-0.005876957438886166,
-0.09072398394346237,
-0.04663916677236557,
-0.024084998294711113,
-0.12404356896877289,
0.08692443370819092,
0.04133107513189316,
0.06254348903894424,
0.10717091709375381,
0.013803672976791859,
-0.026371080428361893,
0.03356083109974861,
0.03029388189315796,
0.04389943927526474,
0.06792975962162018,
0.006386550143361092,
0.13003332912921906,
0.03708616644144058,
0.0026326130609959364,
0.021518070250749588,
0.05598068609833717,
-0.06453469395637512,
-0.05264709144830704,
-0.05363135784864426,
-0.061464227735996246,
0.1591472327709198,
0.11438509076833725,
-0.009315396659076214,
0.102312833070755,
-0.09647942334413528,
0.09166775643825531,
0.1637783944606781,
-0.2882786989212036,
-0.07777463644742966,
0.20046816766262054,
0.07244185358285904,
0.0603218600153923,
-0.07903417944908142,
0.04813934117555618,
0.07434987276792526,
-0.037388820201158524,
0.035469211637973785,
-0.052158061414957047,
-0.006826864089816809,
-0.04611906036734581,
-0.09406235814094543,
0.015696236863732338,
0.2955111265182495,
0.09603045135736465,
-0.06928271055221558,
-0.04375721514225006,
-0.10389520972967148,
-0.10149363428354263,
-0.09432341158390045,
-0.052544090896844864,
0.015783041715621948,
0.008433053269982338,
-0.06447719782590866,
-0.019024508073925972,
-0.08363668620586395,
-0.07629726827144623,
-0.006550298072397709,
0.07046841084957123,
0.02227574773132801,
0.02422613836824894,
-0.04862182214856148,
0.13697046041488647,
0.0012549064122140408,
-0.0734454095363617,
-0.04138766601681709,
-0.003608926199376583,
0.008650701493024826,
-0.03165719285607338,
-0.009743058122694492,
-0.16891233623027802,
0.01227048970758915,
0.04488832876086235,
-0.0386374369263649,
0.04513155668973923,
0.07075220346450806,
0.032552797347307205,
-0.06556641310453415,
0.14096979796886444,
-0.031153604388237,
-0.2000773400068283,
0.07799005508422852,
0.04946637153625488,
0.03693949431180954,
-0.00946301594376564,
-0.09515902400016785,
-0.006685574073344469,
0.055705614387989044,
0.033268705010414124,
-0.048995669931173325,
0.09125696867704391,
0.013004094362258911,
0.010751570574939251,
0.12354283034801483,
-0.06553517282009125,
-0.04471159726381302,
0.004604810383170843,
-0.1329220086336136,
0.0343472994863987,
0.06104952469468117,
-0.02459843084216118,
-0.06418152153491974,
-0.02508021518588066,
-0.053527772426605225,
-0.021879561245441437,
-0.09529341012239456,
-0.09387806057929993,
0.00515956012532115,
0.04550698399543762,
-0.033941008150577545,
-0.1456219106912613,
-0.14273251593112946,
-0.03631294518709183,
0.054385650902986526,
-0.04739232733845711,
-0.03022012673318386,
-0.15539702773094177,
-0.05771606042981148,
0.0403519831597805,
-0.02283121645450592,
0.01892477087676525,
-0.06329506635665894,
0.06383615732192993,
0.040616411715745926,
0.09536448866128922,
-0.06814483553171158,
0.025320494547486305,
-0.07224725931882858,
0.014593919739127159,
-0.21870461106300354,
0.1067652702331543,
-0.0016318962443619967,
0.14994464814662933,
-0.07906203716993332,
-0.05876413360238075,
-0.12493174523115158,
0.030313216149806976,
0.06373237818479538,
0.21343088150024414,
-0.09200994670391083,
-0.03295864537358284,
0.23139920830726624,
-0.030283242464065552,
-0.1102190762758255,
0.16049695014953613,
-0.01074555516242981,
0.16027086973190308,
0.13116271793842316,
0.11933154612779617,
0.01903430186212063,
-0.11287970095872879,
0.0199876856058836,
0.037483736872673035,
-0.007207173854112625,
-0.13981491327285767,
0.024382252246141434,
0.01123865693807602,
-0.022360196337103844,
0.08009656518697739,
-0.0022060817573219538,
0.056389860808849335,
-0.043049585074186325,
-0.019807634875178337,
-0.022671358659863472,
-0.03513320907950401,
-0.05153634399175644,
-0.0020119897089898586,
0.08926927298307419,
-0.14091219007968903,
-0.07999090105295181,
-0.016993500292301178,
0.12068464607000351,
-0.06838250160217285,
-0.009813685901463032,
-0.0762140229344368,
-0.019109798595309258,
-0.05234292894601822,
0.06115783751010895,
-0.034714702516794205,
-0.0435134656727314,
0.003167450660839677,
0.0973883718252182,
0.011100740171968937,
0.15646596252918243,
0.054379403591156006,
0.01370963267982006,
0.020704057067632675,
-0.052640270441770554,
0.03933466225862503,
-0.0399198904633522,
-0.09113630652427673,
-0.05060724541544914,
0.0374276377260685,
-0.05079538747668266,
0.1973830908536911,
-0.1563810110092163,
0.08283516019582748,
0.03182147070765495,
0.015504343435168266,
0.012912847101688385,
-0.004061234649270773,
0.042053624987602234,
-0.04862925782799721,
-0.026485677808523178,
-0.07987677305936813,
0.06080428138375282,
0.09485016018152237,
-0.0036899324040859938,
0.12986144423484802,
-0.16564111411571503,
0.05142557621002197,
0.13412576913833618,
-0.050241243094205856,
-0.04412005841732025,
-0.06066076457500458,
-0.006166127044707537,
-0.043140869587659836,
-0.006141484249383211,
-0.08085456490516663,
0.09375559538602829,
0.03246114403009415,
0.17664559185504913,
-0.0905231162905693,
0.01099070068448782,
-0.0023889662697911263,
-0.06770281493663788,
0.004818136338144541,
0.08326898515224457,
-0.023348484188318253,
-0.16253797709941864,
0.054114654660224915,
0.057336386293172836,
0.029082337394356728,
0.17616768181324005,
0.04566878452897072,
-0.058703403919935226,
-0.013362199068069458,
0.11506751924753189,
-0.022133957594633102,
0.05245967209339142,
-0.25128811597824097,
-0.027076566591858864,
0.011350042186677456,
0.006791837979108095,
0.04617917537689209,
-0.14049258828163147,
0.01897033490240574,
-0.019394338130950928,
-0.05280190706253052,
0.049262385815382004,
0.03790266811847687,
-0.03961031138896942,
0.07904024422168732,
0.027361812070012093,
-0.1902942806482315,
0.04654782637953758,
-0.003280964680016041,
-0.023649724200367928,
0.16076020896434784,
-0.03365068510174751,
-0.2516990602016449,
-0.10005639493465424,
-0.06076589971780777,
-0.06096005439758301,
0.021054083481431007,
0.0625910684466362,
-0.12442559748888016,
-0.06299426406621933,
-0.05378786474466324,
0.032403357326984406,
-0.0057251774705946445,
0.0032010998111218214,
0.0829051062464714,
0.04577309638261795,
-0.016345756128430367,
-0.11865673959255219,
-0.011510411277413368,
-0.05403686314821243,
-0.05294311046600342,
0.05305587500333786,
-0.1420270800590515,
0.05677239969372749,
0.21538659930229187,
-0.03778323903679848,
0.04443467780947685,
0.013059803284704685,
0.13793739676475525,
-0.06194017454981804,
0.03419157862663269,
0.17676325142383575,
0.14767348766326904,
0.026530668139457703,
0.14841985702514648,
0.012669255025684834,
-0.10637661069631577,
0.05291305482387543,
0.002515926957130432,
-0.05779159814119339,
-0.20528171956539154,
-0.06654788553714752,
-0.06600994616746902,
0.11191756278276443,
0.0374378077685833,
0.03303391486406326,
0.04061848670244217,
0.04866795241832733,
-0.06124469265341759,
0.0848776325583458,
0.16778361797332764,
0.06984838843345642,
0.06574870645999908,
-0.03729184716939926,
0.10090930759906769,
-0.048659827560186386,
-0.05919146165251732,
0.04109868034720421,
0.05206020548939705,
0.10728488117456436,
-0.022308828309178352,
0.13430771231651306,
-0.02512682043015957,
0.060994841158390045,
0.012745315209031105,
0.044918645173311234,
0.05213821679353714,
-0.05450952425599098,
-0.05006666108965874,
-0.03321251645684242,
-0.05835788697004318,
0.03417505323886871,
-0.0817909836769104,
-0.031161610037088394,
-0.11722861975431442,
0.11192386597394943,
0.062333784997463226,
0.26671725511550903,
0.11040981858968735,
-0.3114751875400543,
-0.1228172779083252,
-0.0038912841118872166,
0.006622628308832645,
-0.08492685854434967,
0.032275132834911346,
0.09916899353265762,
-0.05618230998516083,
-0.009539317339658737,
-0.05835343897342682,
0.10882797837257385,
-0.06998859345912933,
0.017606498673558235,
-0.058004897087812424,
0.1420648992061615,
-0.043832264840602875,
0.04047444090247154,
-0.17855609953403473,
0.15755583345890045,
0.034262582659721375,
0.12346004694700241,
-0.015530269593000412,
-0.02923930436372757,
0.05717965587973595,
0.1321508139371872,
0.04889899119734764,
-0.02406192384660244,
-0.031165260821580887,
-0.09294691681861877,
-0.16526517271995544,
-0.02582649327814579,
0.02167571522295475,
0.004773914348334074,
0.059125080704689026,
-0.031059952452778816,
-0.017359569668769836,
0.009106965735554695,
-0.06427457183599472,
-0.05768036097288132,
-0.05407996475696564,
0.013642722740769386,
0.18268389999866486,
0.04704318940639496,
-0.08457868546247482,
-0.006240507122129202,
-0.04407396540045738,
0.310245156288147,
0.14592890441417694,
-0.06351874023675919,
-0.07314857840538025,
0.04845494404435158,
0.009256024844944477,
-0.07568231225013733,
0.029195033013820648,
-0.02409297600388527,
0.08840086311101913,
0.014064595103263855,
-0.022483820095658302,
0.1203681081533432,
-0.1007867380976677,
0.009725302457809448,
-0.04032862186431885,
0.13969866931438446,
-0.025102727115154266,
-0.020296446979045868,
0.03364906087517738,
-0.10465719550848007,
-0.06094091758131981,
-0.0521283857524395,
-0.07292438298463821,
0.031210776418447495,
-0.023847658187150955,
0.01782907173037529,
-0.07412490993738174,
-0.0816502645611763,
-0.047314904630184174,
-0.050175171345472336,
0.11210887134075165,
0.1312098652124405,
-0.011687219142913818,
-0.03953912481665611,
0.16673912107944489,
0.04045908898115158,
-0.2510685622692108,
0.0649849995970726,
-0.028564754873514175,
0.003490029601380229,
-0.10056191682815552,
-0.03678416088223457,
0.008533244021236897,
0.06892096996307373,
-0.028362615033984184,
0.14056150615215302,
-0.31368589401245117,
-0.06379471719264984,
0.11667343229055405,
0.04961366206407547,
0.34427890181541443,
-0.09438101202249527,
-0.030487943440675735,
-0.08976733684539795,
-0.16156131029129028,
0.015598558820784092,
-0.14427386224269867,
0.10662143677473068,
0.01159315649420023,
0.10333872586488724,
0.04544037953019142,
-0.07911983877420425,
0.053829263895750046,
-0.05502413213253021,
0.043362054973840714,
-0.09190409630537033,
0.03297301381826401,
0.04245312511920929,
-0.10547809302806854,
0.1807626187801361,
-0.04149443283677101,
0.040082503110170364,
0.04476257413625717,
-0.036523863673210144,
-0.03702162578701973,
0.019751522690057755,
0.013892782852053642,
-0.04285949468612671,
-0.07818092405796051,
0.0027286300901323557,
0.003064299700781703,
0.06106237694621086,
0.10105973482131958,
-0.034924671053886414,
-0.032142523676157,
0.1392928957939148,
-0.00010507254773983732,
-0.2231724113225937,
-0.03875898942351341,
-0.0009558559395372868,
0.018568895757198334,
0.0875403955578804,
-0.12367105484008789,
0.01392298936843872,
0.03761015087366104,
-0.037746723741292953,
0.11433617770671844,
0.02211049571633339,
-0.03294878080487251,
0.028042443096637726,
0.02741716057062149,
-0.08766781538724899,
-0.093525730073452,
-0.0612817257642746,
0.0937749594449997,
-0.03608202934265137,
0.14101997017860413,
0.18407335877418518,
-0.07242739200592041,
0.058191608637571335,
-0.025578735396265984,
0.03156431391835213,
-0.019562989473342896,
0.187423437833786,
-0.03658051788806915,
0.029665665701031685,
-0.09475556761026382,
0.08512040227651596,
0.042817093431949615,
-0.025530101731419563,
-0.01511323917657137,
0.06117820367217064,
-0.16819095611572266,
-0.08003373444080353,
-0.07800329476594925,
0.04201323166489601,
-0.04152153804898262,
-0.04100396856665611,
-0.1787479668855667,
-0.09807970374822617,
-0.001070320256985724,
-0.10256944596767426,
0.015427309088408947,
0.019475966691970825,
-0.014274613000452518,
-0.002939780941233039,
-0.09882449358701706,
0.02737922966480255,
0.05961160734295845,
0.033786576241254807,
-0.09930121898651123,
0.1660037785768509,
0.00487036257982254,
0.022220943123102188,
-0.026368606835603714,
-0.04044167324900627,
-0.0860406905412674,
0.0481550469994545,
-0.045471902936697006,
-0.01646219752728939,
-0.047575656324625015,
-0.021281767636537552,
-0.04136273264884949,
0.03487049788236618,
-0.043630894273519516,
0.020145365968346596,
-0.07211550325155258,
-0.02572648786008358,
-0.015177853405475616,
0.03065498173236847,
-0.006736318580806255,
-0.037725239992141724,
-0.01957138255238533,
-0.08978857100009918,
0.03998788818717003,
0.07528923451900482,
-0.04000254347920418,
-0.021480457857251167,
-0.159033864736557,
0.002951080445200205,
0.11469985544681549,
0.028802646324038506,
0.050828080624341965,
0.04526777192950249,
-0.007614541333168745,
0.07007089257240295,
0.0684470385313034,
-0.021045448258519173,
0.08934243023395538,
-0.06801103055477142,
-0.09139839559793472,
-0.14734932780265808,
-0.040825821459293365,
-0.05127599462866783,
0.04043848440051079,
0.050447337329387665,
0.07846469432115555,
0.14580318331718445,
-0.08207905292510986,
-0.0027730120345950127,
-0.15936170518398285,
-0.01852698065340519,
0.04096319153904915,
-0.1093527153134346,
-0.022818248718976974,
-0.06542560458183289,
0.07753206789493561,
-0.05419338122010231,
0.0271543487906456,
0.0651376023888588,
-0.011555008590221405,
0.007455982733517885,
-0.07515490055084229,
0.052963536232709885,
-0.032309431582689285,
0.14997640252113342,
0.04626503586769104,
0.03441784530878067,
0.009154543280601501,
0.06550638377666473,
0.19418638944625854,
0.051835108548402786,
0.059308961033821106,
0.06712857633829117,
0.007057346869260073,
0.19890622794628143,
-0.075128935277462,
-0.014617808163166046,
-0.12646935880184174,
0.06184782832860947,
-0.06416866183280945,
0.06987743824720383,
-0.04219269007444382,
0.1366613358259201,
0.21217112243175507,
-0.09029071033000946,
-0.0035615984816104174,
-0.009443922899663448,
-0.07262778282165527,
-0.11030013114213943,
-0.07455857843160629,
-0.11359178274869919,
-0.18550634384155273,
-0.002896046033129096,
-0.05499282106757164,
0.03354373201727867,
-0.0461888387799263,
0.029549874365329742,
-0.015472404658794403,
0.18339966237545013,
-0.0575382262468338,
-0.06826373934745789,
0.1300012469291687,
-0.03412824869155884,
-0.0241534560918808,
-0.04567518085241318,
-0.06596212089061737,
0.015603244304656982,
0.028863949701189995,
0.04420212656259537,
-0.013047455810010433,
-0.0025750501081347466,
0.04046336188912392,
0.04789534583687782,
-0.05262007564306259,
-0.06918416172266006,
0.04765242710709572,
0.07279431074857712,
0.10411879420280457,
0.07246262580156326,
-0.08093728125095367,
0.010878106579184532,
0.16920025646686554,
-0.07998662441968918,
-0.010087988339364529,
-0.10197467356920242,
0.021332014352083206,
0.030626045539975166,
0.05165323615074158,
-0.032299816608428955,
-0.060181763023138046,
-0.014148066751658916,
0.1995019018650055,
0.17061682045459747,
-0.06702850759029388,
-0.03372757509350777,
-0.0862005427479744,
-0.013432754203677177,
-0.08162140101194382,
0.1275906264781952,
0.0857699066400528,
0.16383033990859985,
-0.0018952098907902837,
0.003145383670926094,
0.025195792317390442,
-0.041167084127664566,
-0.07564397156238556,
0.034042321145534515,
-0.031544893980026245,
0.015861250460147858,
-0.013747804798185825,
0.06612226366996765,
0.0227273628115654,
-0.13571038842201233,
-0.011850305832922459,
-0.07837969809770584,
-0.04645463451743126,
-0.02964595891535282,
0.0032448433339595795,
0.0002591897500678897,
0.101625956594944,
-0.06785432249307632,
0.045335303992033005,
0.12100126594305038,
-0.04950769618153572,
-0.09628959745168686,
-0.17053498327732086,
0.011918982490897179,
0.037042103707790375,
0.1972997784614563,
-0.023860525339841843,
0.0029841370414942503,
0.09026453644037247,
-0.015197379514575005,
-0.04158936068415642,
0.14832141995429993,
-0.01645125076174736,
-0.016776518896222115,
0.10806632041931152,
-0.0039435094222426414,
-0.08507432043552399,
0.10689493268728256,
0.06551437079906464,
-0.1328834742307663,
-0.007817419245839119,
0.011641053482890129,
-0.0425354540348053,
-0.1401064395904541,
0.053197309374809265,
-0.05012987181544304,
0.07743997126817703,
0.12708523869514465,
-0.06232134997844696,
-0.0017204592004418373,
-0.03834317997097969,
0.0572335347533226,
0.017950598150491714,
-0.03196221962571144,
-0.0032776612788438797,
-0.22093982994556427,
-0.037168215960264206,
-0.0014147425536066294,
0.033721160143613815,
-0.2711721360683441,
-0.03566237911581993,
-0.1018088087439537,
-0.03465652093291283,
-0.06635792553424835,
0.1051696315407753,
0.0947486087679863,
0.019094116985797882,
-0.053087495267391205,
0.02386605739593506,
-0.05185531824827194,
0.1377985030412674,
-0.1167486384510994,
-0.11006075143814087
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | TinyPixel/qwen-1 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-06T05:35:22+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers | # LLaVA Model Card: This is a fork of https://huggingface.co/YouLiXiya/tinyllava-v1.0-1.1b-hf from January 2024

Below is the model card of TinyLlava model 1.1b.
Check out also the Google Colab demo to run Llava on a free-tier Google Colab instance: [](https://colab.research.google.com/drive/1XtdA_UoyNzqiEYVR-iWA-xmit8Y2tKV2#scrollTo=DFVZgElEQk3x)
## Model details
**Model type:**
TinyLLaVA is an open-source chatbot trained by fine-tuning TinyLlama on GPT-generated multimodal instruction-following data.
It is an auto-regressive language model, based on the transformer architecture.
**Paper or resources for more information:**
https://llava-vl.github.io/
## How to use the model
First, make sure to have `transformers >= 4.35.3`.
The model supports multi-image and multi-prompt generation. Meaning that you can pass multiple images in your prompt. Make sure also to follow the correct prompt template (`USER: xxx\nASSISTANT:`) and add the token `<image>` to the location where you want to query images:
### Using `pipeline`:
Below we used [`"YouLiXiya/tinyllava-v1.0-1.1b-hf"`](https://huggingface.co/YouLiXiya/tinyllava-v1.0-1.1b-hf) checkpoint.
```python
from transformers import pipeline
from PIL import Image
import requests
model_id = "YouLiXiya/tinyllava-v1.0-1.1b-hf"
pipe = pipeline("image-to-text", model=model_id)
url = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/ai2d-demo.jpg"
image = Image.open(requests.get(url, stream=True).raw)
prompt = "USER: <image>\nWhat does the label 15 represent? (1) lava (2) core (3) tunnel (4) ash cloud\nASSISTANT:"
outputs = pipe(image, prompt=prompt, generate_kwargs={"max_new_tokens": 200})
print(outputs)
{'generated_text': 'USER: \nWhat does the label 15 represent? (1) lava (2) core (3) tunnel (4) ash cloud\nASSISTANT: The label 15 represents lava, which is the type of rock that is formed from molten magma. '}
```
### Using pure `transformers`:
Below is an example script to run generation in `float16` precision on a GPU device:
```python
import requests
from PIL import Image
import torch
from transformers import AutoProcessor, LlavaForConditionalGeneration
model_id = "YouLiXiya/tinyllava-v1.0-1.1b-hf"
prompt = "USER: <image>\nWhat are these?\nASSISTANT:"
image_file = "http://images.cocodataset.org/val2017/000000039769.jpg"
model = LlavaForConditionalGeneration.from_pretrained(
model_id,
torch_dtype=torch.float16,
low_cpu_mem_usage=True,
).to(0)
processor = AutoProcessor.from_pretrained(model_id)
raw_image = Image.open(requests.get(image_file, stream=True).raw)
inputs = processor(prompt, raw_image, return_tensors='pt').to(0, torch.float16)
output = model.generate(**inputs, max_new_tokens=200, do_sample=False)
print(processor.decode(output[0][2:], skip_special_tokens=True))
```
### Model optimization
#### 4-bit quantization through `bitsandbytes` library
First make sure to install `bitsandbytes`, `pip install bitsandbytes` and make sure to have access to a CUDA compatible GPU device. Simply change the snippet above with:
```diff
model = LlavaForConditionalGeneration.from_pretrained(
model_id,
torch_dtype=torch.float16,
low_cpu_mem_usage=True,
+ load_in_4bit=True
)
```
#### Use Flash-Attention 2 to further speed-up generation
First make sure to install `flash-attn`. Refer to the [original repository of Flash Attention](https://github.com/Dao-AILab/flash-attention) regarding that package installation. Simply change the snippet above with:
```diff
model = LlavaForConditionalGeneration.from_pretrained(
model_id,
torch_dtype=torch.float16,
low_cpu_mem_usage=True,
+ use_flash_attention_2=True
).to(0)
```
## License
Llama 2 is licensed under the LLAMA 2 Community License,
Copyright (c) Meta Platforms, Inc. All Rights Reserved. | {"language": ["en"], "license": "apache-2.0", "pipeline_tag": "image-to-text", "inference": false, "arxiv": 2304.08485} | image-to-text | sujitvasanth/YouLiXiya-tinyllava-v1.0-1.1b-hf | [
"transformers",
"safetensors",
"llava",
"pretraining",
"image-to-text",
"en",
"license:apache-2.0",
"region:us"
] | 2024-02-06T05:46:03+00:00 | [] | [
"en"
] | TAGS
#transformers #safetensors #llava #pretraining #image-to-text #en #license-apache-2.0 #region-us
| # LLaVA Model Card: This is a fork of URL from January 2024
!image/png
Below is the model card of TinyLlava model 1.1b.
Check out also the Google Colab demo to run Llava on a free-tier Google Colab instance:  and add the token '<image>' to the location where you want to query images:
### Using 'pipeline':
Below we used '"YouLiXiya/tinyllava-v1.0-1.1b-hf"' checkpoint.
### Using pure 'transformers':
Below is an example script to run generation in 'float16' precision on a GPU device:
### Model optimization
#### 4-bit quantization through 'bitsandbytes' library
First make sure to install 'bitsandbytes', 'pip install bitsandbytes' and make sure to have access to a CUDA compatible GPU device. Simply change the snippet above with:
#### Use Flash-Attention 2 to further speed-up generation
First make sure to install 'flash-attn'. Refer to the original repository of Flash Attention regarding that package installation. Simply change the snippet above with:
## License
Llama 2 is licensed under the LLAMA 2 Community License,
Copyright (c) Meta Platforms, Inc. All Rights Reserved. | [
"# LLaVA Model Card: This is a fork of URL from January 2024\n\n!image/png\n\n\n\nBelow is the model card of TinyLlava model 1.1b.\n\nCheck out also the Google Colab demo to run Llava on a free-tier Google Colab instance:  and add the token '<image>' to the location where you want to query images:",
"### Using 'pipeline':\n\nBelow we used '\"YouLiXiya/tinyllava-v1.0-1.1b-hf\"' checkpoint.",
"### Using pure 'transformers':\n\nBelow is an example script to run generation in 'float16' precision on a GPU device:",
"### Model optimization",
"#### 4-bit quantization through 'bitsandbytes' library\n\nFirst make sure to install 'bitsandbytes', 'pip install bitsandbytes' and make sure to have access to a CUDA compatible GPU device. Simply change the snippet above with:",
"#### Use Flash-Attention 2 to further speed-up generation\n\nFirst make sure to install 'flash-attn'. Refer to the original repository of Flash Attention regarding that package installation. Simply change the snippet above with:",
"## License\nLlama 2 is licensed under the LLAMA 2 Community License, \nCopyright (c) Meta Platforms, Inc. All Rights Reserved."
] | [
"TAGS\n#transformers #safetensors #llava #pretraining #image-to-text #en #license-apache-2.0 #region-us \n",
"# LLaVA Model Card: This is a fork of URL from January 2024\n\n!image/png\n\n\n\nBelow is the model card of TinyLlava model 1.1b.\n\nCheck out also the Google Colab demo to run Llava on a free-tier Google Colab instance:  and add the token '<image>' to the location where you want to query images:",
"### Using 'pipeline':\n\nBelow we used '\"YouLiXiya/tinyllava-v1.0-1.1b-hf\"' checkpoint.",
"### Using pure 'transformers':\n\nBelow is an example script to run generation in 'float16' precision on a GPU device:",
"### Model optimization",
"#### 4-bit quantization through 'bitsandbytes' library\n\nFirst make sure to install 'bitsandbytes', 'pip install bitsandbytes' and make sure to have access to a CUDA compatible GPU device. Simply change the snippet above with:",
"#### Use Flash-Attention 2 to further speed-up generation\n\nFirst make sure to install 'flash-attn'. Refer to the original repository of Flash Attention regarding that package installation. Simply change the snippet above with:",
"## License\nLlama 2 is licensed under the LLAMA 2 Community License, \nCopyright (c) Meta Platforms, Inc. All Rights Reserved."
] | [
36,
68,
72,
95,
38,
32,
5,
60,
52,
30
] | [
"passage: TAGS\n#transformers #safetensors #llava #pretraining #image-to-text #en #license-apache-2.0 #region-us \n# LLaVA Model Card: This is a fork of URL from January 2024\n\n!image/png\n\n\n\nBelow is the model card of TinyLlava model 1.1b.\n\nCheck out also the Google Colab demo to run Llava on a free-tier Google Colab instance:  and add the token '<image>' to the location where you want to query images:### Using 'pipeline':\n\nBelow we used '\"YouLiXiya/tinyllava-v1.0-1.1b-hf\"' checkpoint.### Using pure 'transformers':\n\nBelow is an example script to run generation in 'float16' precision on a GPU device:### Model optimization#### 4-bit quantization through 'bitsandbytes' library\n\nFirst make sure to install 'bitsandbytes', 'pip install bitsandbytes' and make sure to have access to a CUDA compatible GPU device. Simply change the snippet above with:#### Use Flash-Attention 2 to further speed-up generation\n\nFirst make sure to install 'flash-attn'. Refer to the original repository of Flash Attention regarding that package installation. Simply change the snippet above with:## License\nLlama 2 is licensed under the LLAMA 2 Community License, \nCopyright (c) Meta Platforms, Inc. All Rights Reserved."
] | [
-0.03847626969218254,
0.09233682602643967,
-0.00843183882534504,
0.0368012972176075,
0.11457519978284836,
0.0015363851562142372,
0.1862671971321106,
0.09546951204538345,
0.03192145377397537,
0.08226653933525085,
0.004981920588761568,
0.0680452510714531,
0.08174576610326767,
0.07271385192871094,
0.0887041911482811,
-0.21032311022281647,
0.001327226054854691,
-0.08864390105009079,
0.06000477075576782,
0.05814092233777046,
0.04969799891114235,
-0.0708891823887825,
0.11542053520679474,
0.016312865540385246,
-0.012042884714901447,
-0.004672718234360218,
0.033338017761707306,
-0.019438909366726875,
0.07947055995464325,
0.07294938713312149,
-0.016783909872174263,
0.0379689484834671,
0.04734601452946663,
-0.20309410989284515,
0.02213328517973423,
0.11830092966556549,
-0.018926259130239487,
0.0020929148886352777,
0.08157379180192947,
-0.016668792814016342,
0.16178524494171143,
-0.02524864487349987,
0.02508167363703251,
0.053278665989637375,
-0.08346521854400635,
-0.12101612240076065,
-0.11504661291837692,
0.05369633436203003,
0.12412150949239731,
0.024198148399591446,
0.027132418006658554,
0.05869867652654648,
0.017260583117604256,
0.045555178076028824,
0.14099490642547607,
-0.22241328656673431,
-0.05164625123143196,
0.05189123377203941,
0.03918422758579254,
0.0719301775097847,
-0.04267963021993637,
0.07648036628961563,
-0.009753277525305748,
0.039417095482349396,
0.012008165009319782,
-0.05760583654046059,
0.09448325634002686,
-0.03815339133143425,
-0.08139079809188843,
-0.06929861009120941,
0.07791224867105484,
-0.04739425703883171,
-0.10748332738876343,
-0.12328366190195084,
-0.06224561110138893,
0.055190443992614746,
-0.000848274037707597,
0.05349457636475563,
0.056889310479164124,
0.029545970261096954,
-0.01065603643655777,
-0.13507221639156342,
-0.10960439592599869,
-0.06460180878639221,
-0.01406736858189106,
0.0783650353550911,
0.05788816511631012,
0.06660065799951553,
0.014973537065088749,
0.12448805570602417,
-0.008612442761659622,
-0.038100630044937134,
-0.06743749976158142,
-0.05047448351979256,
-0.06522680819034576,
0.023733489215373993,
-0.014348397962749004,
-0.08663391321897507,
0.06127364560961723,
0.22725483775138855,
0.028795912861824036,
0.027166832238435745,
-0.021787362173199654,
0.02363327518105507,
0.03593805059790611,
0.049321357160806656,
-0.044735223054885864,
0.03372972086071968,
0.030159462243318558,
-0.022076770663261414,
0.11219554394483566,
-0.038870424032211304,
-0.026955705136060715,
0.007728402968496084,
-0.05461324378848076,
0.09638404846191406,
0.06723444163799286,
0.019530991092324257,
-0.008266148157417774,
-0.018165793269872665,
0.12153898179531097,
-0.08640938997268677,
0.04791819676756859,
-0.002522427588701248,
0.002504092874005437,
0.05354496091604233,
0.052499186247587204,
-0.027892619371414185,
-0.06303324550390244,
0.00013179169036448002,
0.0012934874976053834,
0.020660636946558952,
-0.09220394492149353,
-0.019739624112844467,
0.035906847566366196,
-0.038912009447813034,
-0.06824163347482681,
-0.10759734362363815,
-0.16989420354366302,
-0.0012664743699133396,
0.06782027333974838,
-0.029633253812789917,
0.017305919900536537,
0.04642559587955475,
0.021713968366384506,
-0.017578156664967537,
0.00007531628216383979,
-0.005680392496287823,
-0.05325833708047867,
0.0496705025434494,
-0.016546497121453285,
0.053445588797330856,
0.026808202266693115,
0.0029068414587527514,
-0.029517553746700287,
0.0936870276927948,
-0.17464059591293335,
0.060309406369924545,
-0.034610338509082794,
-0.012914631515741348,
-0.05339464545249939,
0.0034161070361733437,
0.05658954754471779,
0.0020002538803964853,
0.014995495788753033,
0.06046513095498085,
-0.1273931860923767,
-0.02077094092965126,
0.13415689766407013,
-0.16581809520721436,
-0.017263226211071014,
0.09584657102823257,
0.024693679064512253,
0.004783570300787687,
0.0829748883843422,
0.04777595028281212,
0.126895010471344,
-0.2196061611175537,
-0.02935861051082611,
0.02646205946803093,
-0.13720835745334625,
-0.03660844638943672,
0.0612313412129879,
-0.014863372780382633,
0.05952155217528343,
0.05278972163796425,
-0.09918665140867233,
0.029459286481142044,
0.005506734363734722,
-0.02430240623652935,
-0.008830353617668152,
-0.023176835849881172,
-0.0660850927233696,
-0.04681447148323059,
-0.07099579274654388,
-0.01829337142407894,
-0.03770602494478226,
-0.002124534919857979,
0.10580030083656311,
-0.05701971426606178,
0.05190180242061615,
-0.08043307811021805,
0.1246657520532608,
-0.03621373698115349,
0.004257199354469776,
-0.03738924488425255,
-0.05540444329380989,
0.05932755023241043,
-0.09765102714300156,
0.03340945020318031,
-0.10906974971294403,
0.02101259119808674,
0.116519495844841,
-0.013755395077168941,
-0.005523747764527798,
0.018196135759353638,
-0.046300262212753296,
0.020664814859628677,
-0.08910449594259262,
-0.04021652042865753,
0.005451316945254803,
0.09717955440282822,
-0.06551508605480194,
0.035861775279045105,
0.08493713289499283,
0.05644381791353226,
0.008859274908900261,
-0.03295324370265007,
0.02356124110519886,
-0.06672843545675278,
-0.009299875237047672,
-0.04805109649896622,
0.05808580666780472,
0.059452466666698456,
0.056085165590047836,
0.11063409596681595,
-0.10585463792085648,
-0.14038363099098206,
0.09901353716850281,
-0.012552103959023952,
-0.02404564991593361,
-0.043432749807834625,
0.006417186465114355,
-0.046189889311790466,
-0.06818096339702606,
-0.029984256252646446,
0.06763160973787308,
0.06916528195142746,
0.07816798239946365,
-0.08072832971811295,
-0.0005868805455975235,
0.030192989856004715,
-0.057508260011672974,
-0.002555428771302104,
0.014310408383607864,
0.13724850118160248,
-0.025323204696178436,
0.036941930651664734,
0.024376777932047844,
0.0044494266621768475,
0.0825953558087349,
0.019804809242486954,
-0.07207167893648148,
-0.057022303342819214,
0.06310895830392838,
0.04986041784286499,
0.10482276231050491,
0.061472684144973755,
0.0020007952116429806,
0.015113655477762222,
-0.043416161090135574,
-0.0060522956773638725,
-0.15758322179317474,
0.02958960086107254,
0.004522477742284536,
-0.04837503656744957,
0.035806041210889816,
-0.014439912512898445,
-0.03767181187868118,
0.054341137409210205,
-0.030653638765215874,
0.11226685345172882,
0.013349156826734543,
-0.02877899631857872,
-0.08093392103910446,
0.10279379785060883,
-0.10059916228055954,
-0.220122829079628,
-0.14757271111011505,
-0.052659641951322556,
-0.11420684307813644,
0.013809961266815662,
0.023186324164271355,
0.02144290693104267,
-0.042742837220430374,
-0.08727717399597168,
-0.016762103885412216,
0.023123281076550484,
-0.07608722150325775,
-0.13542237877845764,
0.0603545606136322,
0.009761831723153591,
-0.10704845190048218,
-0.02151089906692505,
0.033119939267635345,
-0.11530899256467819,
0.07603800296783447,
0.07544732838869095,
0.06935178488492966,
0.026540203019976616,
0.04372444003820419,
-0.015140076167881489,
0.0012028191704303026,
0.09430911391973495,
-0.08051566779613495,
0.08630796521902084,
0.21637751162052155,
0.07325206696987152,
0.10214212536811829,
0.09356805682182312,
0.019793987274169922,
-0.04627666622400284,
-0.000936971278861165,
0.03280670940876007,
-0.08999712765216827,
-0.12381407618522644,
-0.010352658107876778,
-0.008923123590648174,
0.04946625977754593,
0.05975903943181038,
0.07355947047472,
-0.044797006994485855,
0.06325075775384903,
-0.07106591761112213,
-0.02300238236784935,
0.05271320044994354,
0.10085379332304001,
0.13216142356395721,
-0.04394925758242607,
0.039828699082136154,
-0.06049985811114311,
0.04014753922820091,
0.12580296397209167,
0.09313950687646866,
0.14694993197917938,
-0.04233650118112564,
0.09712567925453186,
0.07890774309635162,
0.039351675659418106,
0.00965252984315157,
0.037870559841394424,
-0.012123030610382557,
0.042285896837711334,
-0.009014702402055264,
-0.07773712277412415,
-0.02816721610724926,
0.07450787723064423,
-0.0006282954709604383,
-0.049656812101602554,
0.008990993723273277,
0.0007523542735725641,
0.07764385640621185,
0.24514202773571014,
-0.03599811717867851,
-0.1508018970489502,
0.01633528620004654,
0.008600767701864243,
-0.029773220419883728,
-0.043466124683618546,
0.024933334439992905,
0.1371595859527588,
-0.12972183525562286,
0.06793557107448578,
-0.04224758967757225,
0.04106113687157631,
-0.12860023975372314,
-0.03359675034880638,
0.12201984226703644,
0.11497508734464645,
0.0097612040117383,
0.07201816886663437,
-0.17295759916305542,
-0.03761427104473114,
-0.0037969371769577265,
0.06786762923002243,
-0.016914810985326767,
0.048692841082811356,
0.038539208471775055,
0.02276134490966797,
0.10293012857437134,
0.019190704450011253,
-0.03949125111103058,
-0.08453759551048279,
-0.08110961318016052,
-0.008614784106612206,
0.06569954007863998,
-0.016253620386123657,
0.07284285873174667,
-0.015563363209366798,
-0.026991430670022964,
-0.0803181380033493,
-0.045028023421764374,
-0.10425811260938644,
-0.14842283725738525,
0.03584345430135727,
-0.028414450585842133,
-0.005179538391530514,
-0.09458886086940765,
-0.009956770576536655,
-0.050393495708703995,
0.05943271890282631,
-0.062200646847486496,
-0.11450133472681046,
-0.0777922198176384,
-0.09443427622318268,
0.027039071545004845,
-0.040809229016304016,
0.054201576858758926,
-0.053041718900203705,
0.11253577470779419,
-0.06507561355829239,
-0.062620609998703,
-0.028762144967913628,
-0.0482969805598259,
-0.1721443086862564,
-0.010408761911094189,
0.1244945228099823,
-0.01379100326448679,
-0.019996412098407745,
-0.008879116736352444,
0.03655001148581505,
0.005545508582144976,
-0.114444300532341,
0.027340106666088104,
0.15927055478096008,
0.010173977352678776,
-0.014287146739661694,
-0.03396615758538246,
-0.0029430335853248835,
-0.06363976746797562,
-0.01350076962262392,
0.0552205964922905,
0.21664755046367645,
-0.043708108365535736,
0.1169334352016449,
0.11779430508613586,
-0.10686654597520828,
-0.19900991022586823,
-0.09286008775234222,
-0.014670385047793388,
-0.06213810294866562,
0.0071434807032346725,
-0.20861366391181946,
0.05609379708766937,
0.09094666689634323,
-0.027057012543082237,
0.1305716335773468,
-0.2920204699039459,
-0.09127741307020187,
-0.0833701491355896,
0.13309718668460846,
-0.06494375318288803,
-0.14638182520866394,
-0.05536194145679474,
-0.054774317890405655,
-0.02189687453210354,
0.12499184906482697,
-0.042995385825634,
0.09639938175678253,
-0.00812852568924427,
-0.007534238509833813,
0.015989942476153374,
-0.04067906737327576,
0.08205216377973557,
-0.09206774830818176,
0.07643244415521622,
-0.06460152566432953,
0.028065131977200508,
0.09735811501741409,
-0.05791836977005005,
0.0883818194270134,
-0.11088413000106812,
0.01071833074092865,
-0.0015509825898334384,
-0.0417524017393589,
-0.04677656665444374,
0.05474531650543213,
-0.049884043633937836,
-0.037712328135967255,
-0.030209526419639587,
0.02871663309633732,
-0.0014528256142511964,
-0.010963484644889832,
-0.11822058260440826,
-0.03336578235030174,
-0.14592789113521576,
0.17929604649543762,
0.021401306614279747,
-0.009289922192692757,
-0.13823387026786804,
-0.0661088153719902,
-0.008610056713223457,
0.03842586651444435,
-0.07488982379436493,
0.023787612095475197,
0.02693609707057476,
0.03061814047396183,
0.04478711262345314,
0.01605752855539322,
-0.10958712548017502,
0.027558155357837677,
0.07219716906547546,
-0.1097235232591629,
-0.15489263832569122,
0.005542101804167032,
0.09833206981420517,
0.012925778515636921,
0.007767700590193272,
0.11727055162191391,
-0.041001420468091965,
-0.008502891287207603,
-0.00378590589389205,
0.07496926933526993,
-0.016682103276252747,
0.02926327846944332,
-0.004162817262113094,
-0.03573731705546379,
-0.061906907707452774,
0.13384602963924408,
0.040066782385110855,
-0.08841238915920258,
-0.021172698587179184,
0.12845848500728607,
-0.053749848157167435,
-0.09246905148029327,
-0.10591449588537216,
-0.0007431143894791603,
-0.01420994196087122,
-0.07217857241630554,
0.005266874562948942,
-0.04674485698342323,
0.0038776379078626633,
0.0039058916736394167,
0.023078154772520065,
0.02183922566473484,
0.006355914752930403,
0.02085122838616371,
-0.03823131322860718,
0.10112769156694412,
-0.05394488945603371,
0.04537149518728256,
-0.11115368455648422,
0.04120096191763878,
0.07902354747056961,
0.017324596643447876,
-0.008060604333877563,
-0.06345183402299881,
-0.052236463874578476,
0.02153492346405983,
-0.1082160472869873,
0.06781906634569168,
-0.09380760788917542,
-0.007316868286579847,
0.024941986426711082,
0.04777311906218529,
0.014612934552133083,
0.026003282517194748,
-0.01318227220326662,
-0.05450117588043213,
-0.030958956107497215,
0.10219544172286987,
-0.14802390336990356,
-0.029984010383486748,
0.05773436650633812,
-0.06969994306564331,
0.08135229349136353,
0.006603357382118702,
-0.032876934856176376,
-0.03385242447257042,
-0.03093680366873741,
-0.02446633204817772,
-0.02623257227241993,
0.0438823401927948,
-0.03557756543159485,
-0.1066458448767662,
0.038334622979164124,
0.019016901031136513,
-0.05772266536951065,
-0.036260467022657394,
0.09088023006916046,
-0.1077231764793396,
0.08880192786455154,
0.050219226628541946,
-0.09006872773170471,
-0.06391505897045135,
0.0045132869854569435,
0.03085687942802906,
0.07016713172197342,
0.10346145927906036,
-0.07862377166748047,
0.042703140527009964,
-0.10814160853624344,
-0.02789183333516121,
0.03285479173064232,
-0.000722951372154057,
-0.05811633914709091,
-0.06640732288360596,
0.019724644720554352,
-0.0003433393139857799,
0.13119056820869446,
0.04288860410451889,
-0.032682374119758606,
-0.012906936928629875,
0.025439642369747162,
-0.016288455575704575,
0.030296580865979195,
0.041355229914188385,
-0.012187506072223186,
0.04846207797527313,
-0.011475810781121254,
-0.012156292796134949,
0.007742929272353649,
-0.044997043907642365,
0.019696561619639397,
0.13152284920215607,
0.1254725158214569,
0.04081732779741287,
0.07088874280452728,
-0.05600183084607124,
-0.09253521263599396,
-0.021143203601241112,
-0.0854240357875824,
0.09391558915376663,
-0.09046633541584015,
0.11345606297254562,
0.1166032925248146,
-0.14656779170036316,
0.046189889311790466,
0.03997286036610603,
-0.023199589923024178,
-0.05563902109861374,
-0.2317579686641693,
-0.003379835281521082,
-0.07139910012483597,
0.03261266276240349,
-0.03624512255191803,
0.03508860245347023,
0.09420844167470932,
0.0016808653017506003,
-0.03262843191623688,
0.03331781178712845,
-0.08087257295846939,
-0.0554971843957901,
0.003305230289697647,
0.02285507321357727,
0.04896309971809387,
0.07385314255952835,
-0.006406409200280905,
0.014555206522345543,
-0.01733473315834999,
0.04767792299389839,
0.08891664445400238,
0.115279920399189,
0.053218547254800797,
-0.05100582167506218,
-0.030777819454669952,
0.028223184868693352,
-0.02568541280925274,
0.006281691137701273,
0.08349966257810593,
0.046942681074142456,
-0.021116292104125023,
0.012731272727251053,
0.11961457878351212,
-0.03626028075814247,
-0.07358916103839874,
-0.08475230634212494,
0.1999032199382782,
-0.09353627264499664,
-0.012075298465788364,
-0.030643047764897346,
-0.06359676271677017,
-0.04080750048160553,
0.2234048992395401,
0.1326187551021576,
-0.03410699963569641,
-0.013141871429979801,
-0.005963437259197235,
-0.009150062687695026,
-0.0029131078626960516,
0.15169915556907654,
0.011687999591231346,
0.20926238596439362,
-0.05588012933731079,
0.10263127833604813,
-0.029936881735920906,
-0.012033968232572079,
-0.15341328084468842,
0.092797189950943,
-0.09246599674224854,
0.02701430395245552,
-0.0380871519446373,
0.06392207741737366,
0.03254584223031998,
-0.08453170210123062,
0.09776542335748672,
0.0779130831360817,
-0.015914972871541977,
0.05142270773649216,
0.033705927431583405,
-0.017131764441728592,
0.031040629372000694,
-0.030077453702688217,
0.043325718492269516,
0.15085583925247192,
-0.0008376757032237947,
-0.10097941011190414,
0.007628949359059334,
0.0607723593711853,
-0.09887650609016418,
0.19765540957450867,
-0.00579525763168931,
0.0728016048669815,
0.07105070352554321,
-0.0017631954979151487,
-0.14055418968200684,
0.08696765452623367,
0.017123010009527206,
-0.14620441198349,
-0.0026087563019245863,
0.04251326620578766,
-0.0017464124830439687,
0.035129088908433914,
0.038708996027708054,
0.07483832538127899,
0.045107483863830566,
0.05156569927930832,
0.04254990816116333,
-0.10542024672031403,
0.05032030865550041,
-0.13191558420658112,
0.10259097814559937,
0.06740893423557281,
-0.0011378972558304667,
-0.019502559676766396,
-0.06837157160043716,
0.01811845973134041,
0.03587573021650314,
0.004181364085525274,
0.02702471986413002,
-0.08840285241603851,
0.018408365547657013,
0.005637851543724537,
0.08656276762485504,
-0.13091222941875458,
-0.053961172699928284,
0.011011354625225067,
-0.013585027307271957,
-0.03671326860785484,
0.07681930065155029,
0.0732891783118248,
0.018221629783511162,
-0.04138674959540367,
-0.09602630138397217,
0.0003910423256456852,
0.07078387588262558,
-0.06984522938728333,
-0.07204807549715042
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text2text-generation | Kishan11/nepali-summ | [
"transformers",
"safetensors",
"mt5",
"text2text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T05:47:11+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.051328662782907486,
0.16320262849330902,
-0.005184381268918514,
0.02251756191253662,
0.0970073789358139,
0.014154063537716866,
0.06843415647745132,
0.1107407882809639,
-0.01975325681269169,
0.11416343599557877,
0.03334542736411095,
0.09905029088258743,
0.11230985820293427,
0.15210844576358795,
-0.0026883550453931093,
-0.22916288673877716,
0.04717505723237991,
-0.126259446144104,
-0.0391545407474041,
0.11665356904268265,
0.14936181902885437,
-0.10113005340099335,
0.07722076028585434,
-0.030988991260528564,
-0.0076387980952858925,
-0.032814256846904755,
-0.0584266260266304,
-0.04825444892048836,
0.04778217524290085,
0.07062895596027374,
0.06431572884321213,
0.005459979176521301,
0.09387511014938354,
-0.2666816711425781,
0.019808361306786537,
0.07136410474777222,
-0.0030179128516465425,
0.07510632276535034,
0.060065507888793945,
-0.07584504038095474,
0.09495777636766434,
-0.05221208930015564,
0.1462833732366562,
0.08274403214454651,
-0.09120061993598938,
-0.18708732724189758,
-0.09102199226617813,
0.10069762915372849,
0.182896688580513,
0.048136260360479355,
-0.02258199453353882,
0.09849601238965988,
-0.0877108946442604,
0.012430681847035885,
0.05295773223042488,
-0.0661652609705925,
-0.05250505730509758,
0.06206765025854111,
0.0807318314909935,
0.07389048486948013,
-0.1229020431637764,
-0.02052718587219715,
0.007085064426064491,
0.008346792310476303,
0.08409557491540909,
0.02241266705095768,
0.15073934197425842,
0.03937535360455513,
-0.12801966071128845,
-0.04859030246734619,
0.10540658980607986,
0.04346438869833946,
-0.04617396742105484,
-0.2521538734436035,
-0.02899259887635708,
-0.02667904645204544,
-0.028535552322864532,
-0.03899479657411575,
0.04266434162855148,
-0.008417289704084396,
0.08040466904640198,
-0.008577418513596058,
-0.07472068816423416,
-0.03689853474497795,
0.06301835924386978,
0.05782817304134369,
0.024571044370532036,
-0.012929320335388184,
0.009019458666443825,
0.11621864885091782,
0.10347475856542587,
-0.12458769977092743,
-0.05281980708241463,
-0.06449289619922638,
-0.08029068261384964,
-0.04488494247198105,
0.03654220327734947,
0.03706445172429085,
0.04768291115760803,
0.24621281027793884,
0.01506307628005743,
0.05642811954021454,
0.03732956945896149,
0.009937233291566372,
0.06307817250490189,
0.11257993429899216,
-0.06105290725827217,
-0.10008567571640015,
-0.027280353009700775,
0.08969353139400482,
0.010363527573645115,
-0.03775324299931526,
-0.055724311619997025,
0.06201998516917229,
0.016127094626426697,
0.12256785482168198,
0.09371091425418854,
0.005687997210770845,
-0.06957072764635086,
-0.06580306589603424,
0.1956067681312561,
-0.16264818608760834,
0.04759228602051735,
0.03654341772198677,
-0.041625015437603,
-0.0036778231151401997,
0.011432791128754616,
0.022557828575372696,
-0.02259793132543564,
0.09028376638889313,
-0.054762836545705795,
-0.039149291813373566,
-0.10884293168783188,
-0.03559769690036774,
0.03501499071717262,
0.010106389410793781,
-0.03266524150967598,
-0.032688479870557785,
-0.0845562070608139,
-0.06954096257686615,
0.09501254558563232,
-0.07321515679359436,
-0.05012572929263115,
-0.017229469493031502,
-0.0749860405921936,
0.025271635502576828,
0.020357489585876465,
0.08093643933534622,
-0.021111220121383667,
0.0403398796916008,
-0.05398182198405266,
0.06070630997419357,
0.11319857090711594,
0.03403271734714508,
-0.05239800736308098,
0.061304960399866104,
-0.2453044205904007,
0.09998688101768494,
-0.070001982152462,
0.055851489305496216,
-0.15144193172454834,
-0.02402091957628727,
0.049301061779260635,
0.005924122873693705,
-0.010987705551087856,
0.14017367362976074,
-0.21632637083530426,
-0.026949219405651093,
0.16230235993862152,
-0.09370499104261398,
-0.07971389591693878,
0.05612146109342575,
-0.051730554550886154,
0.1073049008846283,
0.03953683376312256,
-0.027500709518790245,
0.06375151872634888,
-0.1381090134382248,
0.00040212401654571295,
-0.04698554798960686,
-0.020267371088266373,
0.15640391409397125,
0.07842771708965302,
-0.07008927315473557,
0.07522880285978317,
0.023509828373789787,
-0.024433886632323265,
-0.045395378023386,
-0.017890313640236855,
-0.10874514281749725,
0.011851978488266468,
-0.06341767311096191,
0.017641864717006683,
-0.024686299264431,
-0.09168647229671478,
-0.02869567647576332,
-0.17302192747592926,
-0.020798463374376297,
0.08556832373142242,
-0.008557834662497044,
-0.019592439755797386,
-0.11647101491689682,
0.01107026357203722,
0.03302319347858429,
0.004175432957708836,
-0.1334623545408249,
-0.052139077335596085,
0.02637198194861412,
-0.164566308259964,
0.03547604754567146,
-0.058501359075307846,
0.05004860460758209,
0.03211123123764992,
-0.03329542279243469,
-0.02857385016977787,
0.019824394956231117,
0.006224640179425478,
-0.014512771740555763,
-0.24753068387508392,
-0.029559502378106117,
-0.022510329261422157,
0.1669003665447235,
-0.2150150090456009,
0.037831008434295654,
0.071714386343956,
0.15179769694805145,
0.009358161129057407,
-0.03916828706860542,
0.007046530023217201,
-0.07627426087856293,
-0.030073314905166626,
-0.061686236411333084,
-0.007639498449862003,
-0.035809602588415146,
-0.05275571718811989,
0.0509311743080616,
-0.17023569345474243,
-0.03118530660867691,
0.10080438107252121,
0.06444820761680603,
-0.13638457655906677,
-0.017996715381741524,
-0.03701688349246979,
-0.04423511400818825,
-0.05668354406952858,
-0.05839645862579346,
0.10215305536985397,
0.0586850680410862,
0.0463530607521534,
-0.06511835008859634,
-0.07926809787750244,
0.0017478249501436949,
-0.01636536978185177,
-0.024005327373743057,
0.09562139213085175,
0.07809679955244064,
-0.128550186753273,
0.09052938222885132,
0.10546009242534637,
0.08905062079429626,
0.09842319041490555,
-0.021411335095763206,
-0.08413682132959366,
-0.05219675227999687,
0.025027591735124588,
0.01852312870323658,
0.1322891265153885,
-0.011911415494978428,
0.0515037402510643,
0.0410870797932148,
-0.012500960379838943,
0.0133194075897336,
-0.08853328227996826,
0.031934138387441635,
0.034452520310878754,
-0.0191578920930624,
0.03937562555074692,
-0.03862793743610382,
0.02063792385160923,
0.08873654156923294,
0.046638816595077515,
0.04356502741575241,
0.015139794908463955,
-0.04773930460214615,
-0.11410948634147644,
0.16559910774230957,
-0.12778021395206451,
-0.23285086452960968,
-0.14290852844715118,
0.0011466923169791698,
0.033804524689912796,
-0.01184175442904234,
0.0015852059004828334,
-0.0642307698726654,
-0.11860045790672302,
-0.09271383285522461,
0.013562624342739582,
0.046084269881248474,
-0.08686772733926773,
-0.058901336044073105,
0.060286350548267365,
0.04077553004026413,
-0.14541929960250854,
0.019320974126458168,
0.050119441002607346,
-0.09147187322378159,
-0.010142568498849869,
0.08355879038572311,
0.06965820491313934,
0.1800171583890915,
0.01213023066520691,
-0.019780350849032402,
0.03484240919351578,
0.2154977172613144,
-0.13454559445381165,
0.11504307389259338,
0.14155790209770203,
-0.08801047503948212,
0.08181151747703552,
0.19823485612869263,
0.04090484231710434,
-0.10179063677787781,
0.032779160887002945,
0.019162407144904137,
-0.030266085639595985,
-0.24904344975948334,
-0.07182002812623978,
-0.0015384424477815628,
-0.057178061455488205,
0.07619637995958328,
0.08923010528087616,
0.08985473215579987,
0.014106076210737228,
-0.09460747241973877,
-0.08114413172006607,
0.05550093203783035,
0.10361701995134354,
0.013524936512112617,
-0.010953888297080994,
0.08832461386919022,
-0.03412410989403725,
0.021281907334923744,
0.09124857932329178,
0.0012303158873692155,
0.17275752127170563,
0.05731576681137085,
0.18292830884456635,
0.07830806821584702,
0.07006581127643585,
0.010954656638205051,
0.01338646374642849,
0.021367572247982025,
0.02768668904900551,
-0.003376738866791129,
-0.08766022324562073,
-0.012305330485105515,
0.11825862526893616,
0.07214555889368057,
0.015057150274515152,
0.009626545943319798,
-0.03731287643313408,
0.0816538855433464,
0.17225277423858643,
-0.004195837303996086,
-0.18073737621307373,
-0.06175117567181587,
0.0808240994811058,
-0.0942181721329689,
-0.0978553295135498,
-0.02438163384795189,
0.030480818822979927,
-0.17188774049282074,
0.025788409635424614,
-0.01786126010119915,
0.11369330435991287,
-0.13831548392772675,
-0.020397847518324852,
0.06577063351869583,
0.07205205410718918,
0.001269566360861063,
0.05965572968125343,
-0.160665825009346,
0.10748846083879471,
0.014300641603767872,
0.0694030150771141,
-0.09756353497505188,
0.09956212341785431,
-0.003883731784299016,
-0.013259840197861195,
0.13451461493968964,
0.008388550952076912,
-0.0767815113067627,
-0.07929567247629166,
-0.09308768808841705,
-0.010677417740225792,
0.1281345635652542,
-0.1477975994348526,
0.08526764810085297,
-0.032966434955596924,
-0.045358359813690186,
0.0018689122516661882,
-0.10302852839231491,
-0.126983180642128,
-0.18611262738704681,
0.05528288707137108,
-0.13427682220935822,
0.03672889992594719,
-0.10563333332538605,
-0.03443324565887451,
-0.03114200569689274,
0.18988727033138275,
-0.22595620155334473,
-0.06757284700870514,
-0.15271799266338348,
-0.09939435124397278,
0.1442657858133316,
-0.051644183695316315,
0.08306998014450073,
-0.00545003917068243,
0.1809411197900772,
0.0204471405595541,
-0.024139465764164925,
0.0990704819560051,
-0.09589311480522156,
-0.1944846659898758,
-0.0805567279458046,
0.15795263648033142,
0.13470378518104553,
0.03485332801938057,
-0.0028120747301727533,
0.035782139748334885,
-0.017412638291716576,
-0.1234898641705513,
0.022648973390460014,
0.17883644998073578,
0.06719345599412918,
0.024359386414289474,
-0.026545848697423935,
-0.11176556348800659,
-0.06726131588220596,
-0.032448120415210724,
0.03110574185848236,
0.1833028942346573,
-0.0720948874950409,
0.18599410355091095,
0.14474965631961823,
-0.058687858283519745,
-0.19672097265720367,
0.012748281471431255,
0.03311282768845558,
0.004999228753149509,
0.0346582867205143,
-0.20138630270957947,
0.08562208712100983,
0.00026585807790979743,
-0.050549790263175964,
0.1327323019504547,
-0.1709088832139969,
-0.15038524568080902,
0.06999209523200989,
0.037376705557107925,
-0.19945120811462402,
-0.1203133687376976,
-0.09113546460866928,
-0.05348551273345947,
-0.18588998913764954,
0.10221870243549347,
0.028081931173801422,
0.008173973299562931,
0.032946377992630005,
0.027908386662602425,
0.015304015949368477,
-0.0412941575050354,
0.1914987862110138,
-0.02477763034403324,
0.029256334528326988,
-0.08408133685588837,
-0.0700240284204483,
0.04688924551010132,
-0.05717688798904419,
0.07894738763570786,
-0.025546837598085403,
0.012113095261156559,
-0.10591532289981842,
-0.04247612506151199,
-0.031593095511198044,
0.014440414495766163,
-0.09696424007415771,
-0.08737108111381531,
-0.04703705385327339,
0.09430442005395889,
0.0985439270734787,
-0.03574737161397934,
-0.03532914072275162,
-0.07118727266788483,
0.04086564853787422,
0.19116361439228058,
0.17946277558803558,
0.04091198369860649,
-0.07925742864608765,
-0.006378510035574436,
-0.011479404754936695,
0.04257804900407791,
-0.2157202661037445,
0.06462159752845764,
0.050062380731105804,
0.019104523584246635,
0.1179836317896843,
-0.019891055300831795,
-0.15467524528503418,
-0.06981337070465088,
0.06141482666134834,
-0.05946635082364082,
-0.19295614957809448,
0.0024802349507808685,
0.055617205798625946,
-0.1677827537059784,
-0.04761936143040657,
0.0435972660779953,
-0.0037963632494211197,
-0.03886210918426514,
0.018091563135385513,
0.08966289460659027,
0.0023356645833700895,
0.07194776087999344,
0.057763248682022095,
0.08416751027107239,
-0.10276257991790771,
0.0799563005566597,
0.08647435903549194,
-0.08149342983961105,
0.02556447684764862,
0.09818926453590393,
-0.05909668654203415,
-0.031693510711193085,
0.027908844873309135,
0.08353206515312195,
0.015399634838104248,
-0.0418633408844471,
0.011942686513066292,
-0.10034026950597763,
0.06462868303060532,
0.09035695344209671,
0.030880169942975044,
0.013196618296205997,
0.03233819827437401,
0.046119559556245804,
-0.07282152771949768,
0.12195046246051788,
0.029250819236040115,
0.014950193464756012,
-0.0410873182117939,
-0.04673595353960991,
0.02354983240365982,
-0.02780505269765854,
-0.006134995259344578,
-0.0344335101544857,
-0.07343567907810211,
-0.017396869137883186,
-0.1646784543991089,
-0.014960045926272869,
-0.050430599600076675,
0.009284928441047668,
0.027561979368329048,
-0.03680562227964401,
0.0060029830783605576,
0.00910363718867302,
-0.07602906227111816,
-0.06694655120372772,
-0.0237664096057415,
0.09428686648607254,
-0.16284170746803284,
0.022817213088274002,
0.08406320959329605,
-0.11997158825397491,
0.09150046110153198,
0.019169924780726433,
-0.0043096113950014114,
0.025971177965402603,
-0.1505080759525299,
0.03534393385052681,
-0.03375111520290375,
0.013465344905853271,
0.045393940061330795,
-0.22764363884925842,
-0.0006596326129510999,
-0.03479163348674774,
-0.06353507936000824,
-0.009049107320606709,
-0.0390302799642086,
-0.11447589844465256,
0.10432159900665283,
0.00737398024648428,
-0.08732112497091293,
-0.03339998051524162,
0.03389183431863785,
0.08260936290025711,
-0.02571156620979309,
0.15404881536960602,
-0.00210543698631227,
0.0752558782696724,
-0.17060600221157074,
-0.01917494460940361,
-0.008899319916963577,
0.02190987393260002,
-0.01933358423411846,
-0.008870418183505535,
0.04324660077691078,
-0.02540694922208786,
0.18302802741527557,
-0.026055961847305298,
0.023916294798254967,
0.06704080104827881,
0.028971081599593163,
-0.026980608701705933,
0.10390723496675491,
0.05128125473856926,
0.01839378848671913,
0.019961433485150337,
0.007502324413508177,
-0.04053238779306412,
-0.023456495255231857,
-0.1998721808195114,
0.07054254412651062,
0.14529378712177277,
0.09391805529594421,
-0.01637977734208107,
0.08406981825828552,
-0.09878277778625488,
-0.11636010557413101,
0.1169997826218605,
-0.05138581991195679,
-0.002931964583694935,
-0.06808105856180191,
0.12546302378177643,
0.1463906168937683,
-0.19197741150856018,
0.07215752452611923,
-0.06787984073162079,
-0.0486978255212307,
-0.11575321853160858,
-0.19757696986198425,
-0.057408250868320465,
-0.051381710916757584,
-0.015996510162949562,
-0.047199856489896774,
0.07636301219463348,
0.05376458168029785,
0.008772914297878742,
-0.0009363067802041769,
0.06455569714307785,
-0.028812330216169357,
-0.00030995410634204745,
0.02975199744105339,
0.06359979510307312,
0.011518939398229122,
-0.029690392315387726,
0.019902538508176804,
-0.0106467604637146,
0.04097694158554077,
0.0645679235458374,
0.0467936247587204,
-0.030980709940195084,
0.01525089144706726,
-0.036990053951740265,
-0.10704203695058823,
0.041513945907354355,
-0.028119675815105438,
-0.07895249873399734,
0.1515752673149109,
0.023193838074803352,
0.007386222947388887,
-0.020481420680880547,
0.24136120080947876,
-0.07422091066837311,
-0.09643889963626862,
-0.14721740782260895,
0.10216663032770157,
-0.04222400486469269,
0.060273054987192154,
0.04530956223607063,
-0.10375474393367767,
0.01553369965404272,
0.1282518357038498,
0.16521716117858887,
-0.04190836846828461,
0.020468562841415405,
0.028186822310090065,
0.004011107608675957,
-0.036687396466732025,
0.04996533691883087,
0.06798537820577621,
0.15550518035888672,
-0.04796762019395828,
0.09592252224683762,
-0.00007890416600275785,
-0.0971902385354042,
-0.03596784546971321,
0.11750209331512451,
-0.017748000100255013,
0.01799829490482807,
-0.05343281477689743,
0.1203235387802124,
-0.06190110743045807,
-0.2302190214395523,
0.056667957454919815,
-0.06700639426708221,
-0.13796204328536987,
-0.02338138222694397,
0.07993879169225693,
-0.012393412180244923,
0.027088945731520653,
0.07425012439489365,
-0.07398337125778198,
0.19792741537094116,
0.03790948912501335,
-0.056375857442617416,
-0.05510881170630455,
0.08025650680065155,
-0.10229071229696274,
0.27599942684173584,
0.01627948135137558,
0.046000488102436066,
0.10458707809448242,
-0.01374407671391964,
-0.13688722252845764,
0.020522065460681915,
0.09743198752403259,
-0.09572482854127884,
0.04322853684425354,
0.2035953551530838,
-0.0014828984858468175,
0.12146259099245071,
0.0774778351187706,
-0.07723316550254822,
0.04730105772614479,
-0.09168743342161179,
-0.06985042244195938,
-0.09130857139825821,
0.09622733294963837,
-0.07444840669631958,
0.1427851915359497,
0.13178566098213196,
-0.05372144281864166,
0.0102944141253829,
-0.030041957274079323,
0.04689300060272217,
0.003170362673699856,
0.10121745616197586,
0.008502482436597347,
-0.18413890898227692,
0.022574106231331825,
0.013636616058647633,
0.10810644179582596,
-0.16641046106815338,
-0.09936638176441193,
0.044252779334783554,
-0.0010333930840715766,
-0.060718707740306854,
0.12968340516090393,
0.061216048896312714,
0.04452041536569595,
-0.041992466896772385,
-0.026066385209560394,
-0.00883333757519722,
0.13815481960773468,
-0.10480473190546036,
0.0015699166106060147
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# test_2_3136_files
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the audiofolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1824
- Accuracy: 0.9570
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.0927 | 0.96 | 19 | 0.1981 | 0.9347 |
| 0.0582 | 1.97 | 39 | 0.1831 | 0.9506 |
| 0.046 | 2.99 | 59 | 0.1881 | 0.9570 |
| 0.0465 | 4.0 | 79 | 0.3427 | 0.9188 |
| 0.0651 | 4.96 | 98 | 0.2147 | 0.9315 |
| 0.0178 | 5.97 | 118 | 0.1690 | 0.9602 |
| 0.0318 | 6.99 | 138 | 0.2515 | 0.9475 |
| 0.0377 | 8.0 | 158 | 0.1561 | 0.9570 |
| 0.0277 | 8.96 | 177 | 0.2090 | 0.9538 |
| 0.0069 | 9.62 | 190 | 0.1824 | 0.9570 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["audiofolder"], "metrics": ["accuracy"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "test_2_3136_files", "results": [{"task": {"type": "audio-classification", "name": "Audio Classification"}, "dataset": {"name": "audiofolder", "type": "audiofolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9570063694267515, "name": "Accuracy"}]}]}]} | audio-classification | PatricioMN/test_2_3136_files | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"audio-classification",
"generated_from_trainer",
"dataset:audiofolder",
"base_model:facebook/wav2vec2-base",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-06T05:49:25+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us
| test\_2\_3136\_files
====================
This model is a fine-tuned version of facebook/wav2vec2-base on the audiofolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1824
* Accuracy: 0.9570
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
78,
144,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-audiofolder #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.1445913016796112,
0.1444946527481079,
-0.0015851990319788456,
0.06844194233417511,
0.13042736053466797,
0.00700824148952961,
0.111744724214077,
0.11669853329658508,
-0.09496079385280609,
0.1080116555094719,
0.09978131949901581,
0.09602824598550797,
0.05263662338256836,
0.14006486535072327,
-0.03307613357901573,
-0.2804516553878784,
0.011645709164440632,
0.0037666286807507277,
-0.15481172502040863,
0.11576203256845474,
0.08025854825973511,
-0.10575690865516663,
0.06121446192264557,
0.011918771080672741,
-0.13921543955802917,
-0.005010480992496014,
-0.024742133915424347,
-0.06450653076171875,
0.08965495973825455,
0.037107083946466446,
0.08437039703130722,
0.048164017498493195,
0.0934567078948021,
-0.2167752981185913,
0.015245876275002956,
0.08204144239425659,
0.005778573919087648,
0.08975638449192047,
0.10340993851423264,
-0.0218565184623003,
0.08793997764587402,
-0.06980740278959274,
0.06053680181503296,
0.05257406830787659,
-0.09935496747493744,
-0.28923478722572327,
-0.09809301048517227,
0.0776711031794548,
0.1243157833814621,
0.0841943770647049,
-0.030110588297247887,
0.0654609277844429,
-0.0553450807929039,
0.09210129082202911,
0.2497052550315857,
-0.2616812288761139,
-0.06963244080543518,
0.045531656593084335,
0.08247794210910797,
0.0590367317199707,
-0.12306275963783264,
-0.005391140468418598,
0.06538687646389008,
0.019515128806233406,
0.11140917986631393,
0.0042817676439881325,
0.05977768078446388,
-0.011586245149374008,
-0.15008866786956787,
-0.03403743356466293,
0.15313223004341125,
0.1095927432179451,
-0.04826148599386215,
-0.07300195097923279,
-0.029829755425453186,
-0.23380254209041595,
-0.03429574891924858,
0.00029264597105793655,
0.030575865879654884,
-0.053539130836725235,
-0.1176711842417717,
0.027191078290343285,
-0.06657251715660095,
-0.10060005635023117,
0.04404575750231743,
0.11954854428768158,
0.041952915489673615,
-0.01930816099047661,
0.010671298950910568,
0.11625876277685165,
0.03868648409843445,
-0.1612171232700348,
0.001030805753543973,
0.01964445225894451,
-0.0943978801369667,
-0.025521831586956978,
-0.01940908096730709,
-0.011067881248891354,
0.004031598102301359,
0.15852990746498108,
-0.05845949798822403,
0.06207198649644852,
0.04427514597773552,
0.040883634239435196,
-0.06990710645914078,
0.1263851374387741,
-0.08839549869298935,
-0.09787999093532562,
-0.04099050909280777,
0.11216924339532852,
0.007351517211645842,
-0.01891031302511692,
-0.09017963707447052,
0.04514944925904274,
0.10146678239107132,
0.022380825132131577,
-0.018965337425470352,
0.020537905395030975,
-0.07317524403333664,
-0.03377380222082138,
0.05728497356176376,
-0.07793199270963669,
0.03892488032579422,
0.033726125955581665,
-0.06516492366790771,
-0.01699044555425644,
0.01175913866609335,
0.0307893306016922,
0.025204399600625038,
0.1492430567741394,
-0.09014377743005753,
-0.03686206415295601,
-0.0734500139951706,
-0.08596251159906387,
0.040921907871961594,
-0.09006328880786896,
0.022026455029845238,
-0.06415226310491562,
-0.12707136571407318,
-0.04975217953324318,
0.07339464128017426,
-0.039348188787698746,
-0.08411833643913269,
-0.050149496644735336,
-0.10695042461156845,
0.04602337256073952,
-0.024369539692997932,
0.13846930861473083,
-0.06449418514966965,
0.10741672664880753,
0.017905469983816147,
0.0629488080739975,
0.0335141159594059,
0.06922648102045059,
-0.04505513980984688,
0.06683578342199326,
-0.1862550973892212,
0.05147420987486839,
-0.1053059920668602,
0.048738621175289154,
-0.12308508157730103,
-0.10909435153007507,
-0.03321269899606705,
0.012779838405549526,
0.06696681678295135,
0.08758948743343353,
-0.1825111210346222,
-0.11558816581964493,
0.1624344140291214,
-0.08153899013996124,
-0.11962409317493439,
0.11698323488235474,
-0.008478187024593353,
-0.02647448517382145,
0.03785240650177002,
0.15201525390148163,
0.12001927942037582,
-0.11674708873033524,
-0.04744723066687584,
-0.03381362184882164,
0.09500713646411896,
0.002804878633469343,
0.1140999123454094,
-0.028069304302334785,
0.037794094532728195,
-0.010511277243494987,
-0.02894660085439682,
0.040378887206315994,
-0.09030100703239441,
-0.0732143223285675,
-0.015859849750995636,
-0.0962209478020668,
0.030665218830108643,
0.05513467639684677,
0.024698443710803986,
-0.07828884571790695,
-0.14264844357967377,
0.05248900502920151,
0.11658860743045807,
-0.08139751851558685,
0.007046762853860855,
-0.06362234801054001,
0.10153061151504517,
-0.07430316507816315,
-0.03149055317044258,
-0.15540343523025513,
-0.024361221119761467,
0.015934541821479797,
-0.051384180784225464,
0.0021016625687479973,
0.015891218557953835,
0.0644737258553505,
0.07927881926298141,
-0.05885477736592293,
-0.09419186413288116,
-0.06348302215337753,
0.0048876055516302586,
-0.07130929827690125,
-0.25274378061294556,
-0.0784166008234024,
-0.030159005895256996,
0.15557405352592468,
-0.24256937205791473,
0.012541328556835651,
0.028531935065984726,
0.13491646945476532,
0.060560911893844604,
-0.057018082588911057,
-0.001974952407181263,
0.05587482079863548,
-0.022987734526395798,
-0.07442419975996017,
0.02494552917778492,
0.007610581815242767,
-0.09607946872711182,
-0.024409707635641098,
-0.12682212889194489,
0.1530819684267044,
0.10861055552959442,
0.01454928144812584,
-0.08624537289142609,
-0.02566915936768055,
-0.08780433237552643,
-0.05407489836215973,
-0.026556994765996933,
-0.010579762980341911,
0.11555734276771545,
0.02495449036359787,
0.11996063590049744,
-0.09199388325214386,
-0.05062262341380119,
0.057190489023923874,
-0.0027530286461114883,
-0.019397825002670288,
0.12596504390239716,
0.08792481571435928,
-0.06299136579036713,
0.14301040768623352,
0.1283530592918396,
-0.05314169079065323,
0.17630238831043243,
-0.06293290108442307,
-0.11395185440778732,
-0.024893632158637047,
0.010236070491373539,
0.023320723325014114,
0.14680060744285583,
-0.11103150993585587,
0.011254898272454739,
0.016042938455939293,
0.03559233620762825,
0.016186391934752464,
-0.18418064713478088,
-0.014362817630171776,
0.04445416480302811,
-0.05257110670208931,
-0.04138967767357826,
-0.007327836472541094,
-0.019787266850471497,
0.07405146211385727,
0.005663212854415178,
-0.05420086905360222,
0.011436203494668007,
0.0005754734156653285,
-0.07978963106870651,
0.18475958704948425,
-0.0896994024515152,
-0.12821164727210999,
-0.1564999222755432,
-0.01892191357910633,
-0.027905020862817764,
-0.0036174855194985867,
0.05609510838985443,
-0.09478721767663956,
-0.03286289796233177,
-0.052662502974271774,
0.047124143689870834,
-0.02467704378068447,
0.03135497868061066,
0.03636293485760689,
0.024013083428144455,
0.08778389543294907,
-0.1011877954006195,
0.03512893244624138,
-0.009894434362649918,
-0.03312571346759796,
-0.0019677113741636276,
0.03777962923049927,
0.09133739024400711,
0.16136805713176727,
0.050072796642780304,
0.02028528042137623,
-0.024885527789592743,
0.1979094296693802,
-0.1290755718946457,
0.00842217169702053,
0.104881651699543,
-0.021457413211464882,
0.04648618400096893,
0.1475100815296173,
0.05405456945300102,
-0.0822518914937973,
0.019200118258595467,
0.07872502505779266,
-0.029628092423081398,
-0.23875077068805695,
-0.018318241462111473,
-0.041208647191524506,
0.025472188368439674,
0.0865933820605278,
0.03432269021868706,
0.0460808165371418,
0.056915245950222015,
-0.016300130635499954,
0.027889486402273178,
-0.01003263145685196,
0.053718484938144684,
-0.004728627856820822,
0.039629023522138596,
0.1184420958161354,
-0.027857530862092972,
-0.008113687857985497,
0.034529346972703934,
0.013442935422062874,
0.22927924990653992,
0.005040180869400501,
0.1378842145204544,
0.07743123918771744,
0.13648107647895813,
0.01191529631614685,
0.06575489044189453,
-0.000868080707732588,
-0.03671332076191902,
0.015390832908451557,
-0.05711294710636139,
0.005797491408884525,
0.04272177442908287,
0.035352159291505814,
0.05852667987346649,
-0.13263629376888275,
0.03793245553970337,
0.01596393808722496,
0.30579760670661926,
0.08706915378570557,
-0.32382291555404663,
-0.10833267867565155,
0.009086563251912594,
-0.050773464143276215,
-0.04370039701461792,
0.03465893492102623,
0.13812556862831116,
-0.07780043035745621,
0.07698752731084824,
-0.07073862105607986,
0.08980270475149155,
-0.046353619545698166,
-0.00091082404833287,
0.11414021253585815,
0.10205890238285065,
-0.018416468054056168,
0.060428254306316376,
-0.20623402297496796,
0.28096118569374084,
0.009500646963715553,
0.07579272985458374,
-0.0262382160872221,
0.02837071195244789,
0.02429216168820858,
0.019500702619552612,
0.11324910819530487,
-0.009998178109526634,
-0.09042608737945557,
-0.17715419828891754,
-0.1021641194820404,
-0.0033766659907996655,
0.12370351701974869,
-0.08698680996894836,
0.10840018093585968,
-0.023487690836191177,
-0.037158213555812836,
0.057285260409116745,
-0.04906264320015907,
-0.1058715283870697,
-0.09927044808864594,
0.009905575774610043,
-0.0006241665687412024,
0.05104765295982361,
-0.1034843772649765,
-0.1172846257686615,
-0.12191988527774811,
0.135154128074646,
-0.10410413891077042,
-0.015352141112089157,
-0.13319429755210876,
0.08213204145431519,
0.14611345529556274,
-0.0648280680179596,
0.06789349764585495,
0.016132425516843796,
0.13544732332229614,
0.02245350554585457,
-0.028770968317985535,
0.11279522627592087,
-0.09754479676485062,
-0.23059169948101044,
-0.06280352920293808,
0.1777755469083786,
0.05586972460150719,
0.05360095575451851,
-0.03291751444339752,
0.023747127503156662,
0.014739102683961391,
-0.07117065042257309,
0.07567963749170303,
-0.014465748332440853,
0.048461202532052994,
0.022973572835326195,
-0.01746111363172531,
-0.006377226673066616,
-0.027966883033514023,
-0.04808450862765312,
0.09746067970991135,
0.29111066460609436,
-0.08488660305738449,
0.03543788567185402,
0.05922708660364151,
-0.039937473833560944,
-0.15437191724777222,
0.05932604894042015,
0.1396869570016861,
0.03136565908789635,
0.0396195612847805,
-0.1975407898426056,
0.10186803340911865,
0.08735866844654083,
-0.029256949201226234,
0.12463588267564774,
-0.2831665575504303,
-0.12063808739185333,
0.09309408068656921,
0.09234190732240677,
-0.038044098764657974,
-0.15931831300258636,
-0.0709657073020935,
-0.016222629696130753,
-0.15184445679187775,
0.09179434180259705,
-0.09151598811149597,
0.11987946182489395,
0.00817902758717537,
0.011574961245059967,
0.010902808047831059,
-0.05405011400580406,
0.12376608699560165,
0.02411752939224243,
0.07190823554992676,
-0.009272865951061249,
0.019234200939536095,
0.034008562564849854,
-0.06079702451825142,
-0.007043903693556786,
-0.08970135450363159,
0.024918708950281143,
-0.09418358653783798,
-0.0279218889772892,
-0.0823386088013649,
0.0298015084117651,
-0.06491448730230331,
-0.0421450138092041,
-0.024889186024665833,
0.05404825508594513,
0.04989243298768997,
-0.015268267132341862,
0.14257656037807465,
-0.013760262168943882,
0.14554420113563538,
0.11176647990942001,
0.07096070796251297,
-0.004884588997811079,
-0.1222120076417923,
-0.016562342643737793,
-0.027980249375104904,
0.060771647840738297,
-0.137355238199234,
0.03352398797869682,
0.12360254675149918,
0.045035913586616516,
0.13752570748329163,
0.05668340250849724,
-0.08870901167392731,
0.006077019032090902,
0.08332541584968567,
-0.0967627614736557,
-0.10678824782371521,
-0.028371574357151985,
-0.021742111071944237,
-0.1334184855222702,
0.0479256846010685,
0.10027291625738144,
-0.04453162103891373,
0.0041477615013718605,
0.01028007548302412,
0.02329685166478157,
-0.060571812093257904,
0.2118164449930191,
0.059499043971300125,
0.08817080408334732,
-0.09691909700632095,
0.09424956142902374,
0.02989124320447445,
-0.14669887721538544,
0.026392042636871338,
0.051697343587875366,
-0.06021886318922043,
-0.006783455144613981,
0.03512274846434593,
0.10722189396619797,
0.007154790684580803,
-0.06374047696590424,
-0.11941909044981003,
-0.12339504808187485,
0.07165788114070892,
0.13088786602020264,
0.04488794878125191,
0.02714519575238228,
-0.019367579370737076,
0.05543757230043411,
-0.11697739362716675,
0.12975715100765228,
0.08136038482189178,
0.09065305441617966,
-0.21073079109191895,
0.10303746908903122,
0.010129068978130817,
0.0010945219546556473,
-0.010238036513328552,
0.030151162296533585,
-0.11964951455593109,
-0.00219607912003994,
-0.1062735989689827,
-0.03820440545678139,
-0.05940612405538559,
-0.008280539885163307,
-0.007368945982307196,
-0.04710042104125023,
-0.06559715420007706,
0.023666394874453545,
-0.10761014372110367,
-0.04299532249569893,
0.013309898786246777,
0.06356073915958405,
-0.11626272648572922,
-0.0005622629541903734,
0.03754333034157753,
-0.10667885094881058,
0.08193786442279816,
0.03574935346841812,
0.04303416237235069,
0.0204008761793375,
-0.12068643420934677,
0.019765406847000122,
0.051298435777425766,
-0.019208738580346107,
0.04951656609773636,
-0.1577082872390747,
-0.012015855871140957,
-0.04909265786409378,
0.02456493303179741,
-0.013380846939980984,
0.024559736251831055,
-0.12422972172498703,
-0.024866463616490364,
-0.045079395174980164,
-0.03465963900089264,
-0.05544496327638626,
0.04471379145979881,
0.09180222451686859,
0.008473905734717846,
0.19152536988258362,
-0.06527360528707504,
0.029847078025341034,
-0.22804732620716095,
0.007910706102848053,
-0.009895124472677708,
-0.08009957522153854,
-0.06101822853088379,
-0.012847918085753918,
0.06861245632171631,
-0.06358204036951065,
0.08761083334684372,
-0.05521247163414955,
0.03695344552397728,
0.023677559569478035,
-0.08716833591461182,
0.037193670868873596,
0.052313290536403656,
0.18785424530506134,
0.02282416820526123,
-0.014089680276811123,
0.03850360959768295,
0.02098333276808262,
0.08086379617452621,
0.07236603647470474,
0.1582399159669876,
0.1482161432504654,
-0.034856971353292465,
0.0795532613992691,
0.06811272352933884,
-0.12759453058242798,
-0.1780150830745697,
0.11590977013111115,
-0.06515181064605713,
0.1310863047838211,
0.003243641462177038,
0.17241749167442322,
0.10524627566337585,
-0.19797801971435547,
0.040684279054403305,
-0.029819203540682793,
-0.0808611735701561,
-0.10207833349704742,
-0.06917721033096313,
-0.07604768127202988,
-0.19632649421691895,
0.02807401865720749,
-0.11451175808906555,
0.0339348278939724,
0.05906550586223602,
0.023859955370426178,
0.028498094528913498,
0.17931008338928223,
0.016681525856256485,
0.02129758708178997,
0.0747937485575676,
0.027324115857481956,
-0.049487385898828506,
-0.04115169122815132,
-0.0781911164522171,
0.05085401609539986,
-0.04064096137881279,
0.0400019995868206,
-0.08147566020488739,
-0.10110222548246384,
0.08869943767786026,
0.04387277364730835,
-0.09838401526212692,
0.02514687366783619,
-0.00534639460965991,
0.05340495705604553,
0.06708472222089767,
0.006564925890415907,
0.009210173040628433,
-0.020173735916614532,
0.2237674742937088,
-0.09971460700035095,
-0.009482165798544884,
-0.14510799944400787,
0.17550130188465118,
-0.005809358786791563,
-0.0015971953980624676,
0.029524074867367744,
-0.10012141615152359,
0.006913844961673021,
0.16426998376846313,
0.15829791128635406,
-0.013663219287991524,
-0.02528625912964344,
0.025429852306842804,
-0.013991538435220718,
-0.059211261570453644,
0.06169191375374794,
0.11986295133829117,
0.0875503346323967,
-0.06083962321281433,
-0.057770196348428726,
-0.04446359723806381,
-0.05915287509560585,
0.015349640510976315,
0.08644291758537292,
0.034754056483507156,
-0.019837848842144012,
-0.01653372496366501,
0.11061150580644608,
-0.07334767282009125,
-0.1666644662618637,
0.047298017889261246,
-0.16687558591365814,
-0.18443776667118073,
-0.06022214889526367,
0.09057983756065369,
0.02127673104405403,
0.05135536938905716,
0.001745134242810309,
-0.04783912003040314,
0.10648014396429062,
-0.0006236228509806097,
-0.05096658319234848,
-0.10714952647686005,
0.04929694905877113,
-0.0865415558218956,
0.20606544613838196,
-0.05712014436721802,
0.023628951981663704,
0.11554640531539917,
0.05302377790212631,
-0.07260329276323318,
0.023211311548948288,
0.06730476021766663,
-0.13084346055984497,
0.02130172587931156,
0.1911344975233078,
-0.04578228294849396,
0.14574195444583893,
0.03201417624950409,
-0.13560351729393005,
0.010304278694093227,
-0.08288923650979996,
-0.06996539980173111,
-0.06896751374006271,
-0.0032673466484993696,
-0.0492728017270565,
0.13924315571784973,
0.19518136978149414,
-0.06083936616778374,
-0.01432484108954668,
-0.056422580033540726,
0.04369629919528961,
0.07044296711683273,
0.10462351143360138,
0.0005305275553837419,
-0.27808716893196106,
0.03944433480501175,
-0.00007613396883243695,
0.0003337277739774436,
-0.2569994628429413,
-0.07607728987932205,
0.022762393578886986,
-0.06392017006874084,
-0.07881500571966171,
0.07898396998643875,
0.06899280101060867,
0.058612801134586334,
-0.06738528609275818,
-0.02455052174627781,
-0.06532672792673111,
0.17360302805900574,
-0.18616752326488495,
-0.07471124827861786
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SMIDS_5x_beit_large_RMSProp_lr00001_fold2
This model is a fine-tuned version of [microsoft/beit-large-patch16-224](https://huggingface.co/microsoft/beit-large-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.2683
- Accuracy: 0.9085
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.2133 | 1.0 | 750 | 0.4438 | 0.8619 |
| 0.0933 | 2.0 | 1500 | 0.3902 | 0.9068 |
| 0.0066 | 3.0 | 2250 | 0.7047 | 0.8885 |
| 0.0781 | 4.0 | 3000 | 0.7479 | 0.8968 |
| 0.0212 | 5.0 | 3750 | 0.8114 | 0.9052 |
| 0.0005 | 6.0 | 4500 | 0.7637 | 0.9035 |
| 0.095 | 7.0 | 5250 | 0.7892 | 0.8952 |
| 0.0016 | 8.0 | 6000 | 0.7555 | 0.9002 |
| 0.0 | 9.0 | 6750 | 0.8496 | 0.9101 |
| 0.0291 | 10.0 | 7500 | 0.7725 | 0.9218 |
| 0.0031 | 11.0 | 8250 | 1.0613 | 0.8869 |
| 0.0 | 12.0 | 9000 | 0.7920 | 0.9135 |
| 0.0 | 13.0 | 9750 | 1.0124 | 0.9002 |
| 0.0021 | 14.0 | 10500 | 0.9293 | 0.8935 |
| 0.0 | 15.0 | 11250 | 0.8528 | 0.8985 |
| 0.0 | 16.0 | 12000 | 1.0130 | 0.9002 |
| 0.0 | 17.0 | 12750 | 0.8948 | 0.8952 |
| 0.0025 | 18.0 | 13500 | 0.9897 | 0.8952 |
| 0.0 | 19.0 | 14250 | 1.0959 | 0.9002 |
| 0.0 | 20.0 | 15000 | 0.9871 | 0.9151 |
| 0.0 | 21.0 | 15750 | 1.1370 | 0.8968 |
| 0.0 | 22.0 | 16500 | 1.1472 | 0.8935 |
| 0.0032 | 23.0 | 17250 | 0.9326 | 0.9002 |
| 0.0012 | 24.0 | 18000 | 1.1430 | 0.8852 |
| 0.0183 | 25.0 | 18750 | 1.0681 | 0.8985 |
| 0.0 | 26.0 | 19500 | 1.1400 | 0.9052 |
| 0.0136 | 27.0 | 20250 | 1.3202 | 0.8902 |
| 0.0 | 28.0 | 21000 | 1.1445 | 0.8935 |
| 0.0 | 29.0 | 21750 | 1.2039 | 0.8852 |
| 0.0102 | 30.0 | 22500 | 1.0653 | 0.8985 |
| 0.0 | 31.0 | 23250 | 1.0496 | 0.9085 |
| 0.0 | 32.0 | 24000 | 1.0494 | 0.9085 |
| 0.0 | 33.0 | 24750 | 1.2033 | 0.9018 |
| 0.0245 | 34.0 | 25500 | 1.2752 | 0.9052 |
| 0.0048 | 35.0 | 26250 | 1.1112 | 0.9168 |
| 0.0 | 36.0 | 27000 | 1.0978 | 0.9135 |
| 0.0085 | 37.0 | 27750 | 1.3039 | 0.9018 |
| 0.0 | 38.0 | 28500 | 1.1925 | 0.9101 |
| 0.0 | 39.0 | 29250 | 1.2043 | 0.9068 |
| 0.0 | 40.0 | 30000 | 1.1617 | 0.9085 |
| 0.0 | 41.0 | 30750 | 1.2087 | 0.9068 |
| 0.0 | 42.0 | 31500 | 1.2116 | 0.9101 |
| 0.0 | 43.0 | 32250 | 1.2536 | 0.9118 |
| 0.0 | 44.0 | 33000 | 1.2498 | 0.9101 |
| 0.0 | 45.0 | 33750 | 1.2443 | 0.9118 |
| 0.0 | 46.0 | 34500 | 1.2623 | 0.9085 |
| 0.0 | 47.0 | 35250 | 1.2620 | 0.9085 |
| 0.0 | 48.0 | 36000 | 1.2632 | 0.9085 |
| 0.0 | 49.0 | 36750 | 1.2697 | 0.9085 |
| 0.0 | 50.0 | 37500 | 1.2683 | 0.9085 |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1
- Datasets 2.12.0
- Tokenizers 0.13.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/beit-large-patch16-224", "model-index": [{"name": "SMIDS_5x_beit_large_RMSProp_lr00001_fold2", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "test", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.908485856905158, "name": "Accuracy"}]}]}]} | image-classification | onizukal/SMIDS_5x_beit_large_RMSProp_lr00001_fold2 | [
"transformers",
"pytorch",
"beit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/beit-large-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T05:49:49+00:00 | [] | [] | TAGS
#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| SMIDS\_5x\_beit\_large\_RMSProp\_lr00001\_fold2
===============================================
This model is a fine-tuned version of microsoft/beit-large-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.2683
* Accuracy: 0.9085
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.32.1
* Pytorch 2.0.1
* Datasets 2.12.0
* Tokenizers 0.13.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
"TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
81,
116,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
-0.1292150914669037,
0.17132072150707245,
-0.002415567170828581,
0.13183215260505676,
0.11657863855361938,
0.020753253251314163,
0.1335890144109726,
0.16620413959026337,
-0.08238927274942398,
0.04929587244987488,
0.13697229325771332,
0.1357421576976776,
0.04955337569117546,
0.20790311694145203,
-0.053285520523786545,
-0.26080378890037537,
0.0391765721142292,
0.03443576768040657,
-0.020672276616096497,
0.12494900077581406,
0.09484300017356873,
-0.1312379240989685,
0.11272566765546799,
0.025938162580132484,
-0.20840293169021606,
-0.033587437123060226,
-0.01026944164186716,
-0.06854863464832306,
0.10221196711063385,
0.001568986801430583,
0.0741027221083641,
0.037979885935783386,
0.08491890877485275,
-0.12677186727523804,
0.000941311358474195,
0.04326357692480087,
0.0062435888685286045,
0.1065368577837944,
0.062226198613643646,
-0.008521218784153461,
0.06926212459802628,
-0.07453521341085434,
0.06115834787487984,
0.008060229010879993,
-0.11478453874588013,
-0.2692618668079376,
-0.09817449003458023,
0.07377522438764572,
0.08109822124242783,
0.06491127610206604,
0.006432840134948492,
0.16222304105758667,
-0.015434488654136658,
0.1024109497666359,
0.23076069355010986,
-0.2713507413864136,
-0.054792311042547226,
0.022649891674518585,
0.0155020197853446,
0.06252340972423553,
-0.10333037376403809,
-0.01993185468018055,
0.019141921773552895,
0.042880840599536896,
0.14450453221797943,
-0.012332411482930183,
-0.03331032395362854,
-0.02637922763824463,
-0.11139829456806183,
-0.08930420875549316,
0.18604889512062073,
0.06140090152621269,
-0.04917457327246666,
-0.07841385900974274,
-0.07612120360136032,
-0.17419220507144928,
-0.03924720734357834,
0.008911197073757648,
0.046679239720106125,
-0.04711441695690155,
-0.10239296406507492,
-0.03511375933885574,
-0.07504668086767197,
-0.05196268856525421,
-0.026160720735788345,
0.1420334428548813,
0.03879573196172714,
0.05471520125865936,
-0.027205273509025574,
0.10149593651294708,
0.010796112939715385,
-0.1717151701450348,
-0.02661297097802162,
0.0005703883362002671,
0.010487399995326996,
-0.01821139082312584,
-0.029929913580417633,
-0.06737607717514038,
-0.003975129686295986,
0.15347014367580414,
-0.07002666592597961,
0.058850113302469254,
-0.0054583000019192696,
0.041531506925821304,
-0.049319881945848465,
0.1874888390302658,
-0.029916515573859215,
-0.016198426485061646,
0.019476165995001793,
0.08928463608026505,
0.0656052976846695,
-0.030047036707401276,
-0.12371734529733658,
0.021691862493753433,
0.13241209089756012,
0.006458523217588663,
-0.022870952263474464,
0.054544735699892044,
-0.0711979940533638,
-0.0584990456700325,
0.09274657070636749,
-0.09275025129318237,
0.035496871918439865,
-0.011692462489008904,
-0.08981472253799438,
-0.06787234544754028,
0.029122935608029366,
0.011931490153074265,
-0.009771439246833324,
0.06940538436174393,
-0.09093258529901505,
0.01846885494887829,
-0.06650768965482712,
-0.09852384030818939,
0.01388985849916935,
-0.11549968272447586,
0.010918805375695229,
-0.10079170018434525,
-0.19154705107212067,
0.0032797311432659626,
0.07527101784944534,
-0.06246669217944145,
-0.06951755285263062,
-0.033377837389707565,
-0.07729615271091461,
0.03790769353508949,
-0.01523390132933855,
0.07408059388399124,
-0.07056254893541336,
0.09071778506040573,
0.02892814762890339,
0.09002465009689331,
-0.052364569157361984,
0.048610031604766846,
-0.09854818880558014,
0.05158581584692001,
-0.19896768033504486,
0.0824570581316948,
-0.04529954120516777,
0.05730293318629265,
-0.10005063563585281,
-0.10804302245378494,
0.029095064848661423,
-0.0466112419962883,
0.07224688678979874,
0.09985066950321198,
-0.16068536043167114,
-0.05396431311964989,
0.14283035695552826,
-0.09281232208013535,
-0.14269256591796875,
0.09829698503017426,
-0.045770496129989624,
0.014614340849220753,
0.04329100251197815,
0.2130173146724701,
0.04901750758290291,
-0.08417420834302902,
-0.023242823779582977,
-0.02969830296933651,
0.03785223513841629,
-0.0668954998254776,
0.10032020509243011,
0.025215676054358482,
0.05325069651007652,
0.02284027636051178,
-0.029413679614663124,
0.04126512631773949,
-0.08672589063644409,
-0.09880872070789337,
-0.053216658532619476,
-0.0853687971830368,
0.03892384096980095,
0.05334646999835968,
0.0614997074007988,
-0.10279879719018936,
-0.09344549477100372,
0.0453280434012413,
0.09495674818754196,
-0.07567895948886871,
0.02865210548043251,
-0.08989366888999939,
0.10926083475351334,
-0.08635354787111282,
-0.02427433431148529,
-0.18316780030727386,
-0.041861772537231445,
0.04194685444235802,
-0.025394707918167114,
-0.007599220145493746,
-0.05216266214847565,
0.06521623581647873,
0.0848059430718422,
-0.05379978567361832,
-0.05897609516978264,
-0.05670713260769844,
0.002749721286818385,
-0.10883764177560806,
-0.17341645061969757,
-0.08353621512651443,
-0.03381705656647682,
0.14265403151512146,
-0.15880316495895386,
0.019960513338446617,
0.05115775763988495,
0.12808771431446075,
0.060330405831336975,
-0.044940851628780365,
-0.0009795452933758497,
0.02373526245355606,
-0.05278978496789932,
-0.09012233465909958,
0.059676408767700195,
0.0331520177423954,
-0.07579167187213898,
-0.016548609361052513,
-0.09850107133388519,
0.1460651308298111,
0.1280234009027481,
-0.010448831133544445,
-0.04986010119318962,
-0.011923554353415966,
-0.06967874616384506,
-0.030430803075432777,
-0.036602724343538284,
0.019139016047120094,
0.09450183063745499,
0.012393946759402752,
0.14818525314331055,
-0.09332848340272903,
-0.034156475216150284,
0.05024607852101326,
-0.028047295287251472,
-0.03259625658392906,
0.0731319710612297,
0.025664178654551506,
-0.14941470324993134,
0.14837577939033508,
0.14845694601535797,
-0.04714515432715416,
0.12564225494861603,
-0.03889495134353638,
-0.06329566240310669,
-0.04632000997662544,
-0.02844901941716671,
0.013190032914280891,
0.13346467912197113,
-0.076783187687397,
-0.004412572830915451,
0.05686868354678154,
0.017921162769198418,
-0.004722983110696077,
-0.1827412098646164,
0.003951311111450195,
0.0321657620370388,
-0.05121494084596634,
-0.011695281602442265,
-0.017026077955961227,
0.003609517589211464,
0.09151934087276459,
0.02040533348917961,
-0.06441836804151535,
0.05384209007024765,
0.012033452279865742,
-0.05366513133049011,
0.1677880585193634,
-0.07823625206947327,
-0.20364677906036377,
-0.12268579006195068,
-0.06752478331327438,
-0.10258819162845612,
0.012170074507594109,
0.06315170973539352,
-0.04569438472390175,
-0.050954580307006836,
-0.0997823104262352,
-0.037851084023714066,
0.021281057968735695,
0.026625970378518105,
0.05139283835887909,
-0.005415658466517925,
0.09185726940631866,
-0.09241294115781784,
-0.030897676944732666,
-0.01631389558315277,
0.009287231601774693,
0.06772445887327194,
0.019780615344643593,
0.1102219671010971,
0.07713042199611664,
-0.029881305992603302,
0.05137522891163826,
-0.013354548253118992,
0.2620471715927124,
-0.06917091459035873,
-0.002909549279138446,
0.1375615894794464,
-0.015162656083703041,
0.08283410966396332,
0.1273423582315445,
0.041794080287218094,
-0.09746479243040085,
-0.011291430331766605,
-0.0008301159832626581,
-0.049490246921777725,
-0.16143162548542023,
-0.04317644611001015,
-0.0434197373688221,
-0.010716320015490055,
0.1416788250207901,
0.03848205506801605,
0.024626927450299263,
0.07702240347862244,
0.015813151374459267,
0.057987019419670105,
-0.02077260985970497,
0.1017511859536171,
0.0805719867348671,
0.06816057115793228,
0.13305824995040894,
-0.036980245262384415,
-0.02092074789106846,
0.057033997029066086,
0.04002218693494797,
0.21362732350826263,
-0.02804172970354557,
0.15433214604854584,
0.026679744943976402,
0.1909136176109314,
0.019870078191161156,
0.07247955352067947,
-0.010095180943608284,
0.0028269465547055006,
-0.018500015139579773,
-0.04554403945803642,
-0.05979170650243759,
0.03185109794139862,
-0.016015755012631416,
0.05207211896777153,
-0.09269700944423676,
0.028567379340529442,
0.06037893891334534,
0.3028397262096405,
0.061388690024614334,
-0.41139692068099976,
-0.09273239970207214,
0.009406263940036297,
-0.002105827210471034,
-0.06053102761507034,
-0.011343861930072308,
0.09683393687009811,
-0.09968853741884232,
0.08300996571779251,
-0.09414921700954437,
0.08760150521993637,
-0.08863518387079239,
0.016419410705566406,
0.07728815078735352,
0.06722814589738846,
0.01766069419682026,
0.057678405195474625,
-0.22131015360355377,
0.2517315745353699,
0.02006395347416401,
0.04867706075310707,
-0.08515261113643646,
0.013813616707921028,
0.029918700456619263,
0.058915551751852036,
0.08619558066129684,
0.0083828279748559,
-0.09208258241415024,
-0.19043345749378204,
-0.12182265520095825,
-0.0015020827995613217,
0.06677291542291641,
-0.03118232637643814,
0.0942893773317337,
-0.01760665327310562,
-0.012930129654705524,
0.019664883613586426,
0.00020212549134157598,
-0.039232417941093445,
-0.09916181117296219,
0.019594477489590645,
0.03770963475108147,
-0.0040510352700948715,
-0.06473120301961899,
-0.1088499054312706,
-0.027749689295887947,
0.1611177921295166,
0.0489477813243866,
-0.07595206052064896,
-0.14163517951965332,
0.0831608697772026,
0.0844789668917656,
-0.08478974550962448,
0.046326830983161926,
-0.015740465372800827,
0.14427345991134644,
0.02813553437590599,
-0.08791226893663406,
0.10567717254161835,
-0.05589807406067848,
-0.18345315754413605,
-0.035460758954286575,
0.09823724627494812,
0.006449915003031492,
0.047238387167453766,
0.0029976284131407738,
0.05834325775504112,
-0.03208146244287491,
-0.05784951522946358,
0.06896662712097168,
-0.0034485149662941694,
0.1075923964381218,
-0.0061480943113565445,
-0.0032397336326539516,
0.02182089537382126,
-0.04197082296013832,
-0.0014782516518607736,
0.1645156890153885,
0.23995232582092285,
-0.10496784001588821,
0.055536478757858276,
0.030249565839767456,
-0.03645236790180206,
-0.18277540802955627,
0.009984065778553486,
0.08414819091558456,
0.0021475672256201506,
0.040169790387153625,
-0.1663118302822113,
0.05386544391512871,
0.10983236879110336,
-0.04191310703754425,
0.07995743304491043,
-0.2803034782409668,
-0.1190505102276802,
0.08906996995210648,
0.13602600991725922,
0.06884066760540009,
-0.13274545967578888,
-0.045290667563676834,
-0.039063699543476105,
-0.16666166484355927,
0.1351267695426941,
-0.04754851385951042,
0.11997194588184357,
-0.040666740387678146,
0.06989686191082001,
0.015085658058524132,
-0.05448267608880997,
0.14587333798408508,
0.00877679605036974,
0.0857420563697815,
-0.07118549197912216,
0.0021252231672406197,
0.10074540972709656,
-0.0982399731874466,
0.07668103277683258,
-0.08308075368404388,
0.06399426609277725,
-0.11283876746892929,
-0.007322354707866907,
-0.07328318059444427,
0.015542288310825825,
-0.012007588520646095,
-0.043488435447216034,
-0.04113076627254486,
0.03472091257572174,
0.06403200328350067,
-0.015996064990758896,
0.20271754264831543,
0.0629286915063858,
0.08313194662332535,
0.17939580976963043,
0.04974674805998802,
-0.096995510160923,
-0.09814400225877762,
-0.04502987116575241,
-0.028452320024371147,
0.06312472373247147,
-0.13321243226528168,
0.05335186421871185,
0.1209464818239212,
0.008661448024213314,
0.12983813881874084,
0.054849762469530106,
-0.0316605418920517,
0.033173978328704834,
0.06366948038339615,
-0.16513317823410034,
-0.08843576163053513,
-0.011303714476525784,
0.01758752204477787,
-0.12545546889305115,
0.0447046272456646,
0.12079240381717682,
-0.057224519550800323,
-0.015418118797242641,
-0.0026640621945261955,
0.03586487099528313,
-0.00886022113263607,
0.16030296683311462,
0.05005719140172005,
0.05675157532095909,
-0.11541767418384552,
0.1181424930691719,
0.06067226454615593,
-0.0710521712899208,
0.031696248799562454,
0.05698402598500252,
-0.10586927086114883,
-0.022646361961960793,
0.03662630170583725,
0.14154238998889923,
-0.06414706259965897,
-0.04990902543067932,
-0.13196614384651184,
-0.0909038558602333,
0.07024894654750824,
0.0724560096859932,
0.09284354001283646,
0.016252439469099045,
-0.031063025817275047,
-0.014114780351519585,
-0.10623957961797714,
0.10545456409454346,
0.04753988981246948,
0.09451808035373688,
-0.17563696205615997,
0.06374634802341461,
0.0007657874375581741,
0.07206296175718307,
-0.024532334879040718,
0.005616967566311359,
-0.09020458161830902,
-0.0008940583793446422,
-0.10660925507545471,
0.025940274819731712,
-0.04968960955739021,
0.0027822551783174276,
-0.020955873653292656,
-0.058104176074266434,
-0.06385789811611176,
0.02704726532101631,
-0.11796805262565613,
-0.05728267878293991,
0.01832517236471176,
0.029680335894227028,
-0.11609132587909698,
-0.04758497327566147,
0.014494677074253559,
-0.09034118801355362,
0.09993617236614227,
0.05929066613316536,
-0.006737631745636463,
0.0029803363140672445,
0.011042662896215916,
-0.02363271825015545,
0.06827948242425919,
0.006517379079014063,
0.07795335352420807,
-0.11366859823465347,
-0.018052512779831886,
0.017967568710446358,
-0.002112566027790308,
0.011524608358740807,
0.15499049425125122,
-0.12699781358242035,
-0.0033930845092982054,
-0.022802060469985008,
-0.06095515564084053,
-0.06754840165376663,
0.06765563786029816,
0.10613249987363815,
0.0214694757014513,
0.2064255326986313,
-0.054858945310115814,
0.01148067507892847,
-0.21229742467403412,
-0.011367390863597393,
0.0014767643297091126,
-0.1394193321466446,
-0.10240225493907928,
-0.03432944789528847,
0.0646229088306427,
-0.07021024078130722,
0.1212792620062828,
0.036924295127391815,
0.015180133283138275,
0.028698688372969627,
0.025451842695474625,
-0.009322993457317352,
0.01828060857951641,
0.16467928886413574,
0.014544252306222916,
-0.030929861590266228,
0.12307319045066833,
0.026831358671188354,
0.0918813943862915,
0.11550118029117584,
0.17162561416625977,
0.1226300448179245,
0.042329173535108566,
0.09527058154344559,
0.05073356628417969,
-0.032373297959566116,
-0.2198440134525299,
0.04109371080994606,
-0.043747998774051666,
0.14987531304359436,
-0.0034218686632812023,
0.15886609256267548,
0.08696271479129791,
-0.1824999451637268,
0.04266338422894478,
-0.02988567017018795,
-0.08202743530273438,
-0.08238054066896439,
-0.1163601353764534,
-0.10495591163635254,
-0.15148837864398956,
0.0012598474277183414,
-0.10238117724657059,
0.02373862825334072,
0.11528778076171875,
-0.010980993509292603,
-0.00952758826315403,
0.1250862330198288,
-0.01644187793135643,
0.019042596220970154,
0.04508042708039284,
0.007425562012940645,
-0.05218745768070221,
-0.04613304138183594,
-0.08413935452699661,
0.015972480177879333,
0.0363130047917366,
0.05680973082780838,
-0.03208919242024422,
-0.008708061650395393,
0.03847881406545639,
-0.008026620373129845,
-0.12142552435398102,
0.013289375230669975,
0.007551861461251974,
0.04767835885286331,
-0.004989264067262411,
0.007813788950443268,
0.026865217834711075,
-0.01780105195939541,
0.195222407579422,
-0.06977689266204834,
-0.02860948257148266,
-0.12041912227869034,
0.17737813293933868,
0.00569287920370698,
-0.048185933381319046,
0.05394943431019783,
-0.09105358272790909,
-0.02213868498802185,
0.15108588337898254,
0.18787547945976257,
-0.06683575361967087,
-0.017941389232873917,
-0.014669668860733509,
-0.01477136928588152,
-0.01832989603281021,
0.10442051291465759,
0.09986825287342072,
-0.004740583244711161,
-0.07264549285173416,
-0.024389909580349922,
-0.06369390338659286,
-0.032235804945230484,
-0.04127946496009827,
0.07026855647563934,
-0.001124961650930345,
0.005972458980977535,
-0.07571399211883545,
0.03954308480024338,
-0.020357538014650345,
-0.06112333759665489,
0.07204564660787582,
-0.21083933115005493,
-0.1802441030740738,
0.0017737408634275198,
0.07683850824832916,
0.0021866720635443926,
0.04613208398222923,
-0.012570524588227272,
0.018509654328227043,
0.07427240163087845,
-0.02333001233637333,
-0.08794470131397247,
-0.09525144845247269,
0.1020299568772316,
-0.13951729238033295,
0.24700812995433807,
-0.03552914783358574,
0.0377071388065815,
0.1201176866889,
0.03583609312772751,
-0.13580889999866486,
0.03513867408037186,
0.03722600266337395,
-0.02918340638279915,
0.0181744247674942,
0.14616045355796814,
-0.03901152312755585,
0.07440102845430374,
0.04275068640708923,
-0.10678882896900177,
-0.04424819350242615,
-0.04619530588388443,
-0.015570126473903656,
-0.02712010033428669,
-0.05963090807199478,
-0.04089967906475067,
0.12949442863464355,
0.17410574853420258,
-0.04094170406460762,
-0.021948745474219322,
-0.06438223272562027,
0.035308949649333954,
0.08067496865987778,
-0.026465818285942078,
-0.04482371732592583,
-0.2364819198846817,
0.0028874515555799007,
0.050913918763399124,
-0.008316555991768837,
-0.19871793687343597,
-0.10607530176639557,
-0.00044736277777701616,
-0.05943094193935394,
-0.08227076381444931,
0.09325046092271805,
0.06211918964982033,
0.03563893958926201,
-0.06190048158168793,
0.02738066203892231,
-0.07750356942415237,
0.14178979396820068,
-0.14600589871406555,
-0.07656177133321762
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "openai/whisper-large-v3"} | null | kenilshah35/whisper-large-dictation | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:openai/whisper-large-v3",
"region:us"
] | 2024-02-06T05:53:38+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-large-v3 #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-large-v3 #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
40,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-large-v3 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.12347478419542313,
0.2092142105102539,
-0.002684095874428749,
0.03130857273936272,
0.08190006017684937,
0.019184697419404984,
0.048305243253707886,
0.12903062999248505,
0.008469058200716972,
0.10618655383586884,
0.07007666677236557,
0.11070923507213593,
0.11244861781597137,
0.215914785861969,
0.0028417520225048065,
-0.17289768159389496,
0.02935832180082798,
-0.08825898915529251,
0.004292030353099108,
0.12655508518218994,
0.14238038659095764,
-0.1010560393333435,
0.08274201303720474,
-0.013286365196108818,
0.001807458815164864,
-0.03917964920401573,
-0.07005756348371506,
-0.019032185897231102,
0.04317465052008629,
0.03359256312251091,
0.05567975714802742,
-0.009957701899111271,
0.0904567614197731,
-0.25927382707595825,
0.018740130588412285,
0.04293008893728256,
0.004419999662786722,
0.08710487931966782,
0.09760573506355286,
-0.03601525351405144,
0.12154766172170639,
-0.027834249660372734,
0.14213943481445312,
0.09166601300239563,
-0.08594594150781631,
-0.22745291888713837,
-0.06378351151943207,
0.07966628670692444,
0.18843697011470795,
0.07954458892345428,
-0.04141513630747795,
0.13450968265533447,
-0.07146020233631134,
0.025447625666856766,
0.03600752353668213,
-0.08646193146705627,
-0.06822630017995834,
0.06204381212592125,
0.12724728882312775,
0.059036027640104294,
-0.12481880933046341,
-0.03700632229447365,
0.02942323125898838,
0.03816602751612663,
0.06548003107309341,
0.00950158853083849,
0.16484929621219635,
0.025913255289196968,
-0.1432991772890091,
-0.05010582134127617,
0.15236876904964447,
0.019223958253860474,
-0.04536379501223564,
-0.2305048704147339,
-0.006110350601375103,
-0.08914127200841904,
-0.02428293414413929,
-0.05153939500451088,
0.033840522170066833,
0.010660686530172825,
0.12171917408704758,
-0.040523745119571686,
-0.09475444257259369,
-0.022868862375617027,
0.0947757363319397,
0.05170821398496628,
0.023206030949950218,
-0.018395110964775085,
0.011412796564400196,
0.12480875104665756,
0.08659183233976364,
-0.13268519937992096,
-0.06697636842727661,
-0.07995662838220596,
-0.045981112867593765,
-0.037959884852170944,
0.042507968842983246,
0.0347457192838192,
0.06632690876722336,
0.26329314708709717,
-0.020927870646119118,
0.061608947813510895,
0.06735975295305252,
0.01674475520849228,
0.050241876393556595,
0.10413408279418945,
-0.03934614732861519,
-0.16015207767486572,
-0.01081655640155077,
0.09751465171575546,
-0.0034717365633696318,
-0.03103606216609478,
-0.0462617501616478,
0.03457183390855789,
0.036714259535074234,
0.11229129135608673,
0.11473307013511658,
-0.013718659058213234,
-0.07598215341567993,
-0.06453057378530502,
0.21008247137069702,
-0.15897220373153687,
0.044154491275548935,
0.02468547224998474,
-0.007196569349616766,
-0.04750289395451546,
0.007637747097760439,
0.017757810652256012,
-0.028866520151495934,
0.06820913404226303,
-0.06559105962514877,
-0.045095037668943405,
-0.12603087723255157,
-0.023714877665042877,
0.030792664736509323,
0.009655335918068886,
-0.038577184081077576,
-0.040167614817619324,
-0.08166772127151489,
-0.10495708137750626,
0.11010807752609253,
-0.05593036115169525,
-0.0598946288228035,
-0.02923053875565529,
-0.09149757772684097,
0.02420330047607422,
0.027747027575969696,
0.07704121619462967,
-0.027124842628836632,
0.04628084972500801,
-0.01195799745619297,
0.061778854578733444,
0.07831252366304398,
0.028086327016353607,
-0.07878653705120087,
0.06311435252428055,
-0.19246020913124084,
0.08351573348045349,
-0.08278854191303253,
0.03644539788365364,
-0.16099072992801666,
-0.009945225901901722,
0.017655137926340103,
0.025285188108682632,
0.0337202213704586,
0.16664963960647583,
-0.22063106298446655,
-0.023113930597901344,
0.15745165944099426,
-0.10816479474306107,
-0.12850473821163177,
0.040706563740968704,
-0.040591511875391006,
0.17437972128391266,
0.026360297575592995,
0.0022858181037008762,
0.09960731118917465,
-0.16336296498775482,
-0.02786734327673912,
-0.01794264279305935,
-0.003705450100824237,
0.0830402821302414,
0.08945952355861664,
-0.08788390457630157,
0.012299464084208012,
0.014918036758899689,
-0.05991687998175621,
-0.018255092203617096,
-0.039951737970113754,
-0.1069195568561554,
0.006054949481040239,
-0.08751653879880905,
0.021084709092974663,
-0.001968859927728772,
-0.09519893676042557,
-0.007097590249031782,
-0.15598538517951965,
-0.05834999307990074,
0.08947239071130753,
0.0016924722585827112,
-0.0252817552536726,
-0.1094229593873024,
0.05285720154643059,
-0.03420335426926613,
-0.022497139871120453,
-0.13840876519680023,
-0.02131195366382599,
0.019964704290032387,
-0.1429450362920761,
-0.01203920692205429,
-0.11763743311166763,
0.06746350228786469,
0.005774484947323799,
-0.048560306429862976,
-0.04522262513637543,
-0.0015626323875039816,
0.0025470848195254803,
-0.052915915846824646,
-0.2377011477947235,
-0.030966337770223618,
-0.05121353641152382,
0.15288159251213074,
-0.22325648367404938,
0.04320512339472771,
0.03053339570760727,
0.12123231589794159,
0.0027496826369315386,
-0.06722216308116913,
0.022552339360117912,
-0.07450629025697708,
-0.025010570883750916,
-0.07519665360450745,
-0.003956682980060577,
0.0027171967085450888,
-0.03299647569656372,
0.017070593312382698,
-0.11321831494569778,
-0.04619842395186424,
0.09921564906835556,
0.0618419386446476,
-0.15053705871105194,
0.0028408304788172245,
-0.04475977644324303,
-0.06069222837686539,
-0.07592583447694778,
-0.06946105509996414,
0.09008770436048508,
0.055497750639915466,
0.037249546498060226,
-0.07525596022605896,
-0.07397568970918655,
0.0076055643148720264,
-0.02606276422739029,
-0.010877209715545177,
0.11431387811899185,
0.0726604014635086,
-0.10199256986379623,
0.08960752189159393,
0.07049547880887985,
0.027944883331656456,
0.081062451004982,
-0.02979319356381893,
-0.10423033684492111,
-0.030313370749354362,
0.05319220572710037,
0.010205943137407303,
0.17138522863388062,
-0.07296817004680634,
0.05672692880034447,
0.04838613420724869,
-0.04236995428800583,
0.048625096678733826,
-0.08570318669080734,
0.008930697105824947,
0.006301784422248602,
-0.013005251996219158,
0.03003239817917347,
-0.020213337615132332,
0.006469497922807932,
0.07749982178211212,
0.05204914137721062,
0.030717045068740845,
0.02403457649052143,
-0.03729603812098503,
-0.1409980207681656,
0.1806335300207138,
-0.09322783350944519,
-0.23824097216129303,
-0.15962491929531097,
0.059784308075904846,
0.05073394253849983,
-0.014223068952560425,
0.021566614508628845,
-0.0532090961933136,
-0.1061299592256546,
-0.08587963879108429,
-0.00031271655461750925,
0.03448858857154846,
-0.05545566603541374,
-0.06759006530046463,
0.04725148528814316,
0.045004840940237045,
-0.12071169912815094,
0.029226500540971756,
0.06726223975419998,
-0.019843164831399918,
-0.002318392740562558,
0.057782646268606186,
0.09139839559793472,
0.1833011656999588,
-0.003425649832934141,
0.0014301417395472527,
0.06426984816789627,
0.2803294062614441,
-0.15833261609077454,
0.11782047152519226,
0.14148783683776855,
-0.07246032357215881,
0.07369902729988098,
0.18499070405960083,
0.029912695288658142,
-0.09895625710487366,
0.02462814748287201,
0.023857714608311653,
-0.02124759927392006,
-0.26800820231437683,
-0.05619304999709129,
-0.016311032697558403,
-0.08048754185438156,
0.07626232504844666,
0.08878999203443527,
0.0825219601392746,
0.03945625573396683,
-0.06449397653341293,
-0.09950359910726547,
0.02848038636147976,
0.10521212965250015,
-0.018627235665917397,
0.003850172273814678,
0.08136317133903503,
-0.04313458129763603,
0.011067512445151806,
0.09183250367641449,
-0.01806710846722126,
0.14492526650428772,
0.05290010944008827,
0.10105433315038681,
0.08210835605859756,
0.09268852323293686,
-0.007789135444909334,
0.03739666938781738,
0.013862471096217632,
0.026419837027788162,
0.02124408632516861,
-0.08668363094329834,
0.013223384507000446,
0.11189956218004227,
0.03202850744128227,
0.026502706110477448,
0.01859799399971962,
-0.041929539293050766,
0.04321097210049629,
0.1901656538248062,
0.0207973700016737,
-0.20834247767925262,
-0.08550594002008438,
0.05498751997947693,
-0.08116769790649414,
-0.15349648892879486,
-0.010517554357647896,
0.0331919863820076,
-0.16721031069755554,
0.017749950289726257,
-0.03720615431666374,
0.10250581055879593,
-0.09371349960565567,
-0.04139085114002228,
0.11224901676177979,
0.05373401194810867,
-0.01765403524041176,
0.04887305945158005,
-0.18220120668411255,
0.11099335551261902,
0.02899816632270813,
0.07849540561437607,
-0.08957196772098541,
0.10163407027721405,
0.0002422374818706885,
-0.017516965046525,
0.169024258852005,
0.0033265778329223394,
-0.04612065479159355,
-0.07917584478855133,
-0.10107036679983139,
-0.0037599951028823853,
0.08133319765329361,
-0.13469435274600983,
0.07530880719423294,
-0.03150442615151405,
-0.02652093581855297,
-0.008608316071331501,
-0.08407071977853775,
-0.12994098663330078,
-0.1621502935886383,
0.05484431982040405,
-0.09743087738752365,
0.024708718061447144,
-0.0877065435051918,
-0.052789006382226944,
0.008298030123114586,
0.18190208077430725,
-0.22284698486328125,
-0.11035037040710449,
-0.14615564048290253,
-0.11345257610082626,
0.16267304122447968,
-0.040072448551654816,
0.08425810933113098,
0.0009066864149644971,
0.16144679486751556,
0.010296481661498547,
-0.012983142398297787,
0.09132462739944458,
-0.09504955261945724,
-0.1918145716190338,
-0.05290290340781212,
0.16217347979545593,
0.14371928572654724,
0.031401511281728745,
-0.009262526407837868,
0.027899721637368202,
-0.06216268241405487,
-0.12013687938451767,
0.026483720168471336,
0.16725748777389526,
0.06443209946155548,
-0.019468391314148903,
-0.019181791692972183,
-0.10561235249042511,
-0.05570710077881813,
-0.04065164923667908,
-0.009820743463933468,
0.19589629769325256,
-0.06911680102348328,
0.1544889211654663,
0.10912191867828369,
-0.056334588676691055,
-0.21040882170200348,
0.03457631543278694,
0.04732957109808922,
0.022086679935455322,
0.038463618606328964,
-0.18747277557849884,
0.09699133038520813,
-0.012269876897335052,
-0.08188680559396744,
0.16961142420768738,
-0.16748788952827454,
-0.13524359464645386,
0.10579977184534073,
0.024139683693647385,
-0.21752479672431946,
-0.134478360414505,
-0.10040144622325897,
-0.018278535455465317,
-0.13232018053531647,
0.04542160779237747,
0.005838258191943169,
0.005059269722551107,
0.02064664289355278,
0.01002960093319416,
0.034315939992666245,
-0.05303548276424408,
0.21092399954795837,
-0.03448529541492462,
0.00115162692964077,
-0.05126219242811203,
-0.07949978858232498,
0.022732486948370934,
-0.05126079171895981,
0.11484307795763016,
-0.005874676629900932,
0.0331924706697464,
-0.1642684042453766,
-0.04192374274134636,
-0.05143153294920921,
0.0350266769528389,
-0.09130341559648514,
-0.08527211099863052,
-0.04041292890906334,
0.09215570986270905,
0.09726076573133469,
-0.02095925807952881,
-0.005320659372955561,
-0.09095192700624466,
0.06211664900183678,
0.2053217738866806,
0.2024441659450531,
0.06749454885721207,
-0.05435008555650711,
0.021924065425992012,
-0.033864930272102356,
0.04608134925365448,
-0.2140919715166092,
0.04216253384947777,
0.05902514606714249,
0.020635714754462242,
0.06633742898702621,
-0.01036815531551838,
-0.15375939011573792,
-0.07596366107463837,
0.08316674083471298,
-0.0588708370923996,
-0.16758154332637787,
-0.029846778139472008,
0.023590151220560074,
-0.2117733657360077,
-0.04042990878224373,
0.028501132503151894,
-0.01548746693879366,
-0.036714840680360794,
0.020502693951129913,
0.08141309767961502,
-0.02284887060523033,
0.098844975233078,
0.08519003540277481,
0.09375771135091782,
-0.10282693058252335,
0.05928763002157211,
0.07140450924634933,
-0.03883760794997215,
0.03176741674542427,
0.11546622961759567,
-0.045551665127277374,
-0.038572486490011215,
0.07470622658729553,
0.1050533875823021,
0.015916282311081886,
-0.05800483748316765,
0.011717655695974827,
-0.04602072760462761,
0.0579695925116539,
0.09676352143287659,
0.03212767094373703,
0.006808393634855747,
0.06734400987625122,
0.03281727060675621,
-0.08797255158424377,
0.11680296808481216,
0.06036430969834328,
0.017927570268511772,
-0.0564405731856823,
-0.040010880678892136,
-0.01582944579422474,
-0.01431000605225563,
-0.019151082262396812,
-0.0044031087309122086,
-0.08212833106517792,
-0.003769863396883011,
-0.10877107828855515,
0.021967362612485886,
-0.07874654978513718,
0.00748968031257391,
0.03313393145799637,
-0.04780452698469162,
0.0008652046089991927,
0.0008648672956041992,
-0.0728289932012558,
-0.05533535033464432,
-0.012978690676391125,
0.07961034774780273,
-0.13248984515666962,
0.04112215340137482,
0.07691896706819534,
-0.10944104194641113,
0.07049814611673355,
-0.007835596799850464,
0.006490480620414019,
-0.0024662986397743225,
-0.1475415974855423,
0.0552826002240181,
-0.027159767225384712,
-0.006158269941806793,
0.012128972448408604,
-0.19445185363292694,
-0.00860355794429779,
-0.03485282137989998,
-0.06701508164405823,
0.012396075762808323,
-0.009383243508636951,
-0.12012747675180435,
0.10952422022819519,
0.00706335948780179,
-0.05921412259340286,
-0.024290958419442177,
0.03712832182645798,
0.09572794288396835,
-0.010827337391674519,
0.13089650869369507,
-0.02425370179116726,
0.07719992101192474,
-0.17496538162231445,
-0.006438685115426779,
-0.012096010148525238,
0.054261669516563416,
-0.021339181810617447,
-0.03137456253170967,
0.06133636459708214,
-0.023070907220244408,
0.17807234823703766,
-0.008591814897954464,
0.06757612526416779,
0.05148804932832718,
0.011749440804123878,
0.030200395733118057,
0.07834959775209427,
0.06805652379989624,
-0.009525150060653687,
-0.0002561124856583774,
0.04052254930138588,
-0.005474197678267956,
-0.05101078003644943,
-0.16114141047000885,
0.05714362859725952,
0.16148903965950012,
0.05590265616774559,
0.027719460427761078,
0.016363048925995827,
-0.11475326865911484,
-0.0802459865808487,
0.11171130836009979,
-0.02090812660753727,
-0.03205821290612221,
-0.0680278092622757,
0.19220632314682007,
0.13563725352287292,
-0.1969897747039795,
0.06946441531181335,
-0.05307536944746971,
-0.04440882056951523,
-0.1430138200521469,
-0.17921382188796997,
-0.05831427127122879,
-0.04838190972805023,
-0.026701204478740692,
-0.057794395834207535,
0.04713628068566322,
0.034747932106256485,
-0.0005634403205476701,
-0.022716499865055084,
0.10367508977651596,
0.016837138682603836,
-0.025142841041088104,
0.04770026355981827,
0.05846052244305611,
0.03526107966899872,
-0.0908738523721695,
0.007778325583785772,
-0.000753906846512109,
0.021972278133034706,
0.0700388103723526,
0.02027447707951069,
-0.06757032126188278,
0.02511257864534855,
-0.02055417001247406,
-0.12235992401838303,
0.039766550064086914,
-0.013212153688073158,
-0.03741868585348129,
0.14669184386730194,
0.039743728935718536,
0.009684635326266289,
-0.018208637833595276,
0.2275155931711197,
-0.08102208375930786,
-0.0779198631644249,
-0.14827804267406464,
0.055876970291137695,
-0.07428983598947525,
0.026949143037199974,
0.030589217320084572,
-0.121348075568676,
0.011981568299233913,
0.16982300579547882,
0.12177624553442001,
-0.011504197493195534,
0.009477870538830757,
0.04816402494907379,
0.004543608985841274,
-0.043202366679906845,
0.01931837387382984,
0.04638928547501564,
0.18644310534000397,
-0.07312262058258057,
0.05966896563768387,
-0.014200511388480663,
-0.08376814424991608,
-0.015887608751654625,
0.09291340410709381,
-0.00861785002052784,
-0.0006889643264003098,
-0.06455893069505692,
0.14736807346343994,
-0.08461539447307587,
-0.20973917841911316,
0.06033577397465706,
-0.05673464760184288,
-0.1375177651643753,
-0.04054504632949829,
0.0343482680618763,
-0.020611394196748734,
0.004032730124890804,
0.06895170360803604,
-0.045484695583581924,
0.18657957017421722,
0.03124731034040451,
-0.0485793873667717,
-0.08142749965190887,
0.057663824409246445,
-0.15171681344509125,
0.28401660919189453,
0.02300601825118065,
0.05649800971150398,
0.10847774147987366,
-0.019032655283808708,
-0.1512838751077652,
0.006872016470879316,
0.10915490984916687,
-0.07134553045034409,
0.064127117395401,
0.1672511100769043,
0.007295482326298952,
0.1250617504119873,
0.06161846965551376,
-0.0508042573928833,
0.03313418850302696,
-0.08456829190254211,
-0.04398053511977196,
-0.12108675390481949,
0.0807415023446083,
-0.08905551582574844,
0.15919984877109528,
0.11588563024997711,
-0.07165335863828659,
0.0011541851563379169,
-0.019058609381318092,
0.08632516115903854,
0.011345315724611282,
0.10710826516151428,
0.011513849720358849,
-0.19115038216114044,
0.036095596849918365,
0.007114403415471315,
0.10184800624847412,
-0.18427206575870514,
-0.05228700116276741,
0.0416451096534729,
-0.020221775397658348,
-0.07534758001565933,
0.11600546538829803,
0.03458233177661896,
0.028921110555529594,
-0.037660498172044754,
-0.027667667716741562,
0.009954401291906834,
0.14656680822372437,
-0.11270771920681,
-0.015248586423695087
] |
null | null | transformers | # miqu-1-120b

* EXL2: 2.4bpw | [2.65bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2) | [3.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2) | [4.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2) | [5.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2)
* GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/) | [IQ3_XXS](https://huggingface.co/wolfram/miqu-1-120b-GGUF)
* HF FP16: [wolfram/miqu-1-120b](https://huggingface.co/wolfram/miqu-1-120b)
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
Inspired by [Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2), [MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [goliath-120b](https://huggingface.co/alpindale/goliath-120b).
Thanks for the support, [CopilotKit](https://github.com/CopilotKit/CopilotKit) - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, [Lone Striker](https://huggingface.co/LoneStriker)!
## Prompt template: Mistral
```
<s>[INST] {prompt} [/INST]
```
See also: [🐺🐦⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates : LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/18ljvxb/llm_prompt_format_comparisontest_mixtral_8x7b/)
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
dtype: float16
merge_method: passthrough
slices:
- sources:
- layer_range: [0, 20]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [10, 30]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [20, 40]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [30, 50]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [40, 60]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [50, 70]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [60, 80]
model: 152334H/miqu-1-70b-sf
```
## Credits & Special Thanks
* original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
* leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
* f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
* mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
* mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
### Support
* [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| {"language": ["en", "de", "fr", "es", "it"], "library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["152334H/miqu-1-70b-sf"]} | text-generation | LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"en",
"de",
"fr",
"es",
"it",
"base_model:152334H/miqu-1-70b-sf",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T05:54:05+00:00 | [] | [
"en",
"de",
"fr",
"es",
"it"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # miqu-1-120b
!image/jpeg
* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw
* GGUF: Q2_K-Q5_K_M | IQ3_XXS
* HF FP16: wolfram/miqu-1-120b
This is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.
Inspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.
Thanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, Lone Striker!
## Prompt template: Mistral
See also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* 152334H/miqu-1-70b-sf
### Configuration
The following YAML configuration was used to produce this model:
## Credits & Special Thanks
* original (unreleased) model: mistralai (Mistral AI_)
* leaked model: miqudev/miqu-1-70b
* f16 model: 152334H/miqu-1-70b-sf
* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.
* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2
### Support
* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| [
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
85,
206,
44,
31,
4,
17,
28,
17,
107,
69,
43
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140## Merge Details### Merge Method\n\nThis model was merged using the passthrough merge method.### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.08452296257019043,
0.07996358722448349,
-0.006216804962605238,
0.07489742338657379,
0.062447689473629,
0.048032015562057495,
0.10850238054990768,
0.11730548739433289,
0.04997159540653229,
0.1304655373096466,
0.019059622660279274,
0.0882313996553421,
0.10314993560314178,
0.1546645164489746,
-0.0008784975507296622,
-0.13700617849826813,
0.03406155854463577,
-0.032812073826789856,
-0.05265757441520691,
0.07489487528800964,
0.07002948969602585,
-0.049538981169462204,
0.07827795296907425,
0.01992727816104889,
-0.0794241651892662,
-0.02351442351937294,
-0.03350690379738808,
-0.0034194772597402334,
0.06382977962493896,
0.09272785484790802,
0.010276054963469505,
0.01770416460931301,
0.010751184076070786,
-0.1683615893125534,
0.032938648015260696,
0.03592948243021965,
-0.017932593822479248,
0.07718798518180847,
0.10788709670305252,
-0.018908003345131874,
0.021088771522045135,
-0.10027661174535751,
-0.02568136155605316,
0.08249933272600174,
-0.10719147324562073,
-0.1474289745092392,
-0.18762712180614471,
0.15491966903209686,
0.042507048696279526,
0.03724713996052742,
-0.011902958154678345,
0.1333865076303482,
0.050249699503183365,
0.06016452983021736,
0.24593202769756317,
-0.22131793200969696,
-0.061919860541820526,
0.028957664966583252,
0.05541115999221802,
-0.04985284432768822,
-0.021356018260121346,
0.01992829702794552,
0.04054830223321915,
0.02191905304789543,
-0.020568734034895897,
-0.04234932363033295,
0.13632027804851532,
-0.05843184515833855,
-0.13171182572841644,
-0.02820662222802639,
0.09477535635232925,
0.04355064406991005,
-0.060202740132808685,
-0.1276080310344696,
-0.0682656466960907,
-0.03270931541919708,
-0.0022935839369893074,
-0.04131745919585228,
-0.002290143631398678,
-0.011816466227173805,
0.08523242175579071,
-0.08333887904882431,
-0.05337991565465927,
-0.020235886797308922,
-0.06682160496711731,
0.16135647892951965,
0.018354052677750587,
0.013679140247404575,
-0.013255320489406586,
0.07744071632623672,
-0.1265559196472168,
-0.1336318850517273,
-0.05428304150700569,
-0.03462158143520355,
-0.05692019686102867,
-0.05218517407774925,
0.01661340333521366,
-0.07750479131937027,
0.08107020705938339,
0.175857275724411,
-0.11741548776626587,
0.06220667064189911,
-0.009697522968053818,
0.030903007835149765,
0.024629943072795868,
0.08424429595470428,
-0.08733733743429184,
-0.08122188597917557,
0.04335138201713562,
0.05078619346022606,
0.05994807183742523,
0.0019238290842622519,
0.00010135882621398196,
-0.02575274370610714,
0.0019080113852396607,
0.07156599313020706,
0.06952427327632904,
0.029285242781043053,
-0.07679720222949982,
-0.05144278332591057,
0.17693321406841278,
-0.10877041518688202,
0.05875063315033913,
0.038095518946647644,
-0.010931058786809444,
-0.015534471720457077,
0.026578398421406746,
-0.028613878414034843,
-0.05422482639551163,
0.014574683271348476,
-0.052837468683719635,
0.006087095942348242,
-0.04121243208646774,
-0.04512706398963928,
0.052960656583309174,
0.044456131756305695,
-0.05848181992769241,
-0.12249575555324554,
-0.08548451960086823,
-0.056048691272735596,
0.033982738852500916,
-0.07165446132421494,
0.009947382844984531,
0.03666377440094948,
-0.08962082862854004,
0.029198521748185158,
0.02603435143828392,
0.022042686119675636,
-0.022196292877197266,
0.013299040496349335,
0.01655365526676178,
0.02530047670006752,
-0.031476203352212906,
-0.009267413057386875,
-0.05801313742995262,
0.08978775143623352,
-0.19187140464782715,
0.07319818437099457,
-0.06753759831190109,
-0.0458916574716568,
-0.1367833912372589,
-0.022929007187485695,
0.0001557070208946243,
-0.021685365587472916,
0.08853033930063248,
0.14028236269950867,
-0.15254990756511688,
-0.01800709031522274,
0.133183553814888,
-0.09939759224653244,
-0.08947071433067322,
0.09990517795085907,
0.010765177197754383,
-0.005259351339191198,
0.010271748527884483,
0.08585681021213531,
0.12536770105361938,
-0.056073881685733795,
-0.0745566338300705,
-0.10543492436408997,
-0.02353603020310402,
0.10377749055624008,
0.036487940698862076,
-0.06648851186037064,
0.006608173716813326,
0.0013413212727755308,
-0.03926616534590721,
-0.02470051497220993,
-0.01713505946099758,
-0.04516416788101196,
-0.051653604954481125,
-0.03216031193733215,
0.019755378365516663,
-0.0311356820166111,
-0.04306983947753906,
-0.06874691694974899,
-0.0662546455860138,
-0.03937284275889397,
0.10621660202741623,
0.0002958594122901559,
0.013465981930494308,
-0.06993132084608078,
0.12132570147514343,
-0.043282829225063324,
0.019509179517626762,
-0.13229045271873474,
-0.062326252460479736,
0.05427157133817673,
-0.17642687261104584,
0.057915039360523224,
-0.07802051305770874,
0.07007356733083725,
0.08792263269424438,
-0.03142979368567467,
-0.04443249851465225,
0.014312747865915298,
-0.011503690853714943,
-0.05503445118665695,
-0.13744132220745087,
-0.05339666083455086,
-0.03718583658337593,
0.1360798180103302,
-0.07717297226190567,
0.03823299705982208,
0.04849797114729881,
0.17298123240470886,
0.013948954641819,
-0.06230805069208145,
0.059276606887578964,
0.013416923582553864,
-0.01901276968419552,
-0.04254152253270149,
0.022224929183721542,
-0.015195275656878948,
-0.06335578858852386,
0.06797880679368973,
-0.16682805120944977,
-0.09226492792367935,
0.08205856382846832,
0.11543388664722443,
-0.10825490206480026,
-0.0011853284668177366,
-0.0016989412251859903,
-0.03144387900829315,
-0.02349192649126053,
-0.07283974438905716,
0.139337420463562,
0.03313908725976944,
0.05388158559799194,
-0.05538925528526306,
-0.07884406298398972,
0.012225146405398846,
-0.024691954255104065,
-0.07743193954229355,
0.12690365314483643,
0.02291349321603775,
-0.19036413729190826,
0.1086549460887909,
0.10671419650316238,
0.024346476420760155,
0.12972186505794525,
0.010163333266973495,
-0.020192109048366547,
-0.11323220282793045,
0.07214321941137314,
0.06328365206718445,
0.08764441311359406,
0.009260048158466816,
0.06931837648153305,
0.03807743266224861,
0.00207220995798707,
0.013925828970968723,
-0.11039725691080093,
0.03326145187020302,
0.019972439855337143,
-0.035299137234687805,
0.07873526960611343,
0.04182936251163483,
0.027880189940333366,
0.08705969154834747,
0.017179058864712715,
0.07423700392246246,
-0.01186748594045639,
-0.03507009521126747,
-0.09978814423084259,
0.13680647313594818,
-0.10421305149793625,
-0.1235826313495636,
-0.15689323842525482,
0.023495908826589584,
-0.1045532375574112,
-0.031237754970788956,
0.04314960539340973,
-0.04964744672179222,
-0.04190392047166824,
-0.09863945841789246,
0.07575149089097977,
0.04664899781346321,
-0.031183557584881783,
0.029061678797006607,
-0.04268911853432655,
0.05956493318080902,
-0.1079893708229065,
-0.03650837019085884,
0.0311584435403347,
0.003986986353993416,
0.03826166316866875,
0.06534901261329651,
0.09445244818925858,
0.06816873699426651,
0.010970757342875004,
-0.007111722603440285,
0.0249604694545269,
0.2451404184103012,
-0.09102492034435272,
0.08363116532564163,
0.1731153130531311,
0.012367261573672295,
0.08963974565267563,
0.1418784111738205,
0.0519220232963562,
-0.06685937941074371,
-0.013835576362907887,
0.023638667538762093,
-0.024145755916833878,
-0.19078323245048523,
-0.061981018632650375,
-0.049455784261226654,
0.02179974503815174,
0.016594912856817245,
0.056652046740055084,
-0.03246590495109558,
0.04957956075668335,
-0.06632956862449646,
-0.023145724087953568,
0.04272812604904175,
0.07885605841875076,
0.04942423850297928,
0.05232582241296768,
0.032079048454761505,
-0.030630147084593773,
-0.01874661073088646,
0.10800144821405411,
0.002024435205385089,
0.0828050896525383,
0.005763496737927198,
0.15540826320648193,
0.042188387364149094,
0.05763785168528557,
-0.006694222800433636,
0.05031910166144371,
0.012587900273501873,
0.01727849431335926,
0.011400274932384491,
-0.0922514945268631,
0.017045721411705017,
0.037224024534225464,
0.031022431328892708,
0.08605832606554031,
-0.019477766007184982,
-0.042653512209653854,
0.06828916072845459,
0.20636118948459625,
0.054764457046985626,
-0.18104644119739532,
-0.052608687430620193,
0.05323893949389458,
-0.005172972101718187,
-0.0441189780831337,
-0.0574423223733902,
0.0601525716483593,
-0.1421785205602646,
0.10297753661870956,
-0.02837306633591652,
0.08344317972660065,
-0.0768112912774086,
-0.025202374905347824,
-0.002551008015871048,
0.14314106106758118,
-0.007685830816626549,
0.009870217181742191,
-0.11473074555397034,
0.09677925705909729,
0.047673650085926056,
0.050363969057798386,
-0.0467650331556797,
0.05014784261584282,
0.06751372665166855,
-0.016121957451105118,
0.1009698212146759,
0.015580941922962666,
-0.08597638458013535,
-0.13121242821216583,
-0.07546716183423996,
-0.02328302152454853,
0.09724711626768112,
-0.07956860214471817,
0.11797893047332764,
-0.006677214056253433,
-0.041914381086826324,
-0.051285892724990845,
0.08619783818721771,
-0.16726025938987732,
-0.10657192021608353,
0.08049735426902771,
0.035933516919612885,
-0.00293326354585588,
-0.07889305055141449,
-0.03560107201337814,
-0.06792481988668442,
0.20780132710933685,
-0.12493043392896652,
-0.06458897888660431,
-0.12241725623607635,
-0.02887725457549095,
0.14838698506355286,
-0.09519194811582565,
0.016095273196697235,
-0.0620892271399498,
0.09675031155347824,
-0.05750473216176033,
-0.09322291612625122,
0.03633593022823334,
-0.0772782415151596,
-0.22163733839988708,
-0.029905345290899277,
0.1838245987892151,
0.031322285532951355,
0.03489852324128151,
0.0206821970641613,
0.04595774784684181,
0.021382097154855728,
-0.10996323078870773,
0.04105592891573906,
0.12817256152629852,
0.020900849252939224,
0.09039270132780075,
-0.045222893357276917,
-0.14004118740558624,
-0.09013182669878006,
-0.05677556246519089,
0.0997968390583992,
0.29861730337142944,
-0.05223260074853897,
0.030688846483826637,
0.04147891700267792,
-0.07043230533599854,
-0.15785279870033264,
-0.038915667682886124,
0.04657349735498428,
0.01753225363790989,
0.01858132891356945,
-0.07110119611024857,
0.0366257019340992,
0.10373353213071823,
-0.01873767003417015,
0.15667074918746948,
-0.263369619846344,
-0.12704867124557495,
-0.006682985462248325,
0.06700930744409561,
-0.01957916095852852,
-0.1452975571155548,
-0.11674544960260391,
-0.03796976059675217,
-0.1995268017053604,
0.09981389343738556,
-0.02467156946659088,
0.0965239554643631,
-0.023898916319012642,
0.015778150409460068,
0.03433018922805786,
-0.059150923043489456,
0.19815395772457123,
-0.039659541100263596,
0.04237253591418266,
-0.08535932749509811,
0.011777906678617,
0.026407601311802864,
-0.05703800171613693,
0.0950261577963829,
-0.035693515092134476,
0.017066331580281258,
-0.07961391657590866,
0.0030619590543210506,
-0.08434901386499405,
0.06988205015659332,
-0.053291499614715576,
-0.002292127348482609,
-0.06203543022274971,
0.07322024554014206,
0.02764309197664261,
-0.02415998838841915,
0.024202430620789528,
-0.021172653883695602,
0.10087262094020844,
0.2350863218307495,
0.09775305539369583,
0.05048970505595207,
-0.06141071021556854,
-0.00026774543221108615,
-0.04018312692642212,
0.02579716220498085,
-0.035933900624513626,
0.04462524875998497,
0.08787034451961517,
0.02455063909292221,
0.10782863199710846,
0.01810472458600998,
-0.1601009964942932,
0.018065448850393295,
0.06646080315113068,
-0.13265873491764069,
-0.2303391844034195,
0.003492377931252122,
0.04154038056731224,
-0.05837193876504898,
0.01806606538593769,
0.1862204670906067,
0.0009367137681692839,
-0.04963487386703491,
0.022825637832283974,
0.05416732281446457,
-0.03536570817232132,
0.15667866170406342,
0.00837104581296444,
0.04538920149207115,
-0.061773963272571564,
0.07580430060625076,
0.04042122885584831,
0.0016077554319053888,
0.018714869394898415,
0.17067234218120575,
-0.05190492421388626,
-0.07857772707939148,
0.04102415591478348,
0.03600002080202103,
-0.021241383627057076,
-0.021828800439834595,
-0.027304736897349358,
-0.10624261945486069,
0.013062734156847,
0.12654060125350952,
0.02681851200759411,
-0.018991822376847267,
0.08323623985052109,
-0.004140008706599474,
-0.028763674199581146,
0.07244459539651871,
0.05491362512111664,
0.08497325330972672,
-0.059819940477609634,
0.06284529715776443,
-0.02697847969830036,
0.03915136307477951,
-0.0026654857210814953,
0.019256316125392914,
-0.12592482566833496,
-0.05442740023136139,
-0.08961838483810425,
0.005896841175854206,
-0.11454344540834427,
-0.029031215235590935,
-0.0207052081823349,
-0.00936585757881403,
-0.02115669846534729,
0.03387326002120972,
-0.051367443054914474,
-0.11562348157167435,
-0.07470326870679855,
0.08342844247817993,
-0.14418748021125793,
-0.01641484536230564,
0.06919168680906296,
-0.07516539096832275,
0.09250383824110031,
0.01708120107650757,
0.0034909669775515795,
-0.009268661960959435,
-0.11126343160867691,
-0.02617836929857731,
-0.013749503530561924,
0.020847713574767113,
0.02727784588932991,
-0.20117603242397308,
-0.002813884522765875,
-0.04270019009709358,
-0.03368119150400162,
0.01241485308855772,
0.15425018966197968,
-0.09736622869968414,
-0.040856342762708664,
-0.03006616234779358,
-0.044825147837400436,
-0.06249745562672615,
0.018799860030412674,
0.10122360289096832,
0.009257921017706394,
0.08748460561037064,
-0.05421772226691246,
0.07078760117292404,
-0.17954121530056,
-0.01305952575057745,
0.022268623113632202,
-0.06352261453866959,
0.042893052101135254,
0.006506424397230148,
0.07432542741298676,
-0.024414660409092903,
0.033989157527685165,
-0.08212705701589584,
-0.018906600773334503,
0.026504697278141975,
-0.028520310297608376,
-0.021283207461237907,
0.017214683815836906,
0.020455483347177505,
-0.022489050403237343,
-0.024124853312969208,
-0.0313778817653656,
0.020932979881763458,
-0.019497908651828766,
-0.05568772181868553,
0.1307651400566101,
0.10116901993751526,
0.042394958436489105,
0.06033344939351082,
0.0699099600315094,
-0.06935380399227142,
-0.019917812198400497,
-0.015662092715501785,
-0.0501786470413208,
0.05921950191259384,
-0.0570748895406723,
0.11598431318998337,
0.0971740260720253,
-0.16128812730312347,
0.08817395567893982,
-0.03033575415611267,
-0.025268414989113808,
-0.0654015764594078,
-0.1747601330280304,
-0.06647072732448578,
-0.021373050287365913,
-0.007110625971108675,
-0.07304833084344864,
0.04277768358588219,
0.03683268278837204,
0.0017412106972187757,
-0.0061371480114758015,
0.12685923278331757,
-0.04393889009952545,
-0.058910902589559555,
0.035390980541706085,
0.037289973348379135,
-0.015837162733078003,
0.04472239688038826,
-0.006923268549144268,
0.022451741620898247,
0.012978330254554749,
0.05661128833889961,
0.08223818242549896,
-0.00805328506976366,
0.02584826387465,
-0.028271563351154327,
-0.10848145186901093,
-0.0006025367183610797,
0.005287060514092445,
0.01833254098892212,
0.08337931334972382,
0.05019523203372955,
0.016249405220150948,
-0.015350534580647945,
0.0810377299785614,
-0.02552863396704197,
-0.014691580086946487,
-0.1009829118847847,
0.15841826796531677,
-0.0559045635163784,
0.002254345454275608,
-0.0015080315060913563,
-0.11343378573656082,
-0.011597293429076672,
0.12920117378234863,
0.16337698698043823,
-0.08000685274600983,
0.03376281261444092,
0.017094949260354042,
0.00689410325139761,
0.004397694021463394,
0.049244899302721024,
0.059573933482170105,
0.10981784015893936,
-0.046494387090206146,
0.12055498361587524,
-0.0517595037817955,
-0.032632842659950256,
-0.07947799563407898,
0.09397511184215546,
-0.047008223831653595,
0.014640634879469872,
-0.017156826332211494,
0.07035721093416214,
-0.032913386821746826,
-0.16172361373901367,
0.03033449873328209,
-0.12312323600053787,
-0.138545423746109,
-0.00011208858632016927,
-0.0077992831356823444,
0.001197756384499371,
0.07809285819530487,
0.009989638812839985,
0.010767842642962933,
0.13262617588043213,
0.004294542595744133,
-0.07419104129076004,
-0.056784018874168396,
0.0005215644487179816,
-0.09722831845283508,
0.18984216451644897,
0.024254996329545975,
0.04178498685359955,
0.13091936707496643,
-0.033293675631284714,
-0.15407414734363556,
0.05585271492600441,
0.0700296014547348,
-0.09110260009765625,
0.022162333130836487,
0.14707012474536896,
-0.005258307326585054,
0.030615337193012238,
0.07387614995241165,
-0.053392838686704636,
0.019069518893957138,
0.0722138062119484,
0.019377747550606728,
-0.10239636152982712,
0.08977871388196945,
-0.10703950375318527,
0.13002605736255646,
0.18838851153850555,
-0.03489691764116287,
0.012865101918578148,
-0.03789073973894119,
0.04828639328479767,
0.038723692297935486,
0.09065895527601242,
-0.034608691930770874,
-0.19317959249019623,
0.06994378566741943,
-0.023887865245342255,
0.08475272357463837,
-0.1766246110200882,
-0.1239243745803833,
-0.029406068846583366,
-0.021863967180252075,
-0.013726417906582355,
0.0984341949224472,
0.09909792244434357,
0.0073255919851362705,
-0.023081263527274132,
-0.1434156447649002,
-0.02458539977669716,
0.10146303474903107,
-0.08631070703268051,
-0.04105154797434807
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-base-lora-1.77M-snli-model2
This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7454
- Accuracy: 0.729
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 256
- eval_batch_size: 256
- seed: 27
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.5225 | 1.0 | 2146 | 0.4209 | 0.8470 |
| 0.4716 | 2.0 | 4292 | 0.3864 | 0.8562 |
| 0.4577 | 3.0 | 6438 | 0.3797 | 0.8591 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "t5-base", "model-index": [{"name": "t5-base-lora-1.77M-snli-model2", "results": []}]} | text-classification | varun-v-rao/t5-base-lora-1.77M-snli-model2 | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text-classification",
"generated_from_trainer",
"base_model:t5-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T05:56:37+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text-classification #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| t5-base-lora-1.77M-snli-model2
==============================
This model is a fine-tuned version of t5-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7454
* Accuracy: 0.729
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 256
* eval\_batch\_size: 256
* seed: 27
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 27\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text-classification #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 27\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
74,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text-classification #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 27\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.08444473147392273,
0.09445285052061081,
-0.002206584205850959,
0.10973017662763596,
0.14101476967334747,
0.0062677632085978985,
0.16297617554664612,
0.11735495924949646,
-0.07650014758110046,
0.04575206711888313,
0.12562991678714752,
0.12411639839410782,
0.030785663053393364,
0.13894754648208618,
-0.08589791506528854,
-0.22002537548542023,
0.013962888158857822,
0.020629338920116425,
-0.05578804016113281,
0.12833809852600098,
0.09693676978349686,
-0.11155229806900024,
0.09785313904285431,
-0.026787377893924713,
-0.1471806764602661,
0.009528528898954391,
0.022975699976086617,
-0.06765798479318619,
0.1349010318517685,
0.04769524186849594,
0.11395692825317383,
0.029412707313895226,
0.0673680305480957,
-0.20686796307563782,
0.013666767627000809,
0.07010907679796219,
-0.012503226287662983,
0.08537737280130386,
0.04322627931833267,
-0.00043857263517566025,
0.09454509615898132,
-0.1050054058432579,
0.053072165697813034,
0.024661356583237648,
-0.1329835206270218,
-0.18663200736045837,
-0.07118090987205505,
0.03696947172284126,
0.09918078035116196,
0.08676064014434814,
-0.018193842843174934,
0.13988222181797028,
-0.033540308475494385,
0.10504327714443207,
0.2079431116580963,
-0.32357093691825867,
-0.06184225529432297,
0.04440436512231827,
0.02960634045302868,
0.10705279558897018,
-0.09380437433719635,
-0.007291721645742655,
0.058205585926771164,
0.029070932418107986,
0.1368325799703598,
-0.024282533675432205,
-0.05429835990071297,
0.0006316539365798235,
-0.14335514605045319,
-0.036258913576602936,
0.17758607864379883,
0.060791146010160446,
-0.04519859328866005,
-0.059583310037851334,
-0.07829838991165161,
-0.13969945907592773,
-0.031013930216431618,
-0.0051779551431536674,
0.047390446066856384,
-0.012555547058582306,
-0.050135087221860886,
-0.021546615287661552,
-0.10663975030183792,
-0.06491120159626007,
-0.05086566135287285,
0.11308116465806961,
0.03324295952916145,
-0.0005048795719631016,
-0.0322108194231987,
0.0946829617023468,
-0.014811878092586994,
-0.13807618618011475,
0.02105035074055195,
0.02412416599690914,
0.010873193852603436,
-0.04653508961200714,
-0.04389895871281624,
-0.09718579798936844,
0.028668340295553207,
0.11776310205459595,
-0.06786147505044937,
0.059402190148830414,
-0.019269878044724464,
0.04787922278046608,
-0.10439742356538773,
0.15771029889583588,
-0.034595388919115067,
-0.03297395259141922,
0.030638139694929123,
0.09064029157161713,
0.06716570258140564,
-0.016903726384043694,
-0.1333373337984085,
0.02582559920847416,
0.1236489787697792,
0.03714451938867569,
-0.04989233240485191,
0.08181833475828171,
-0.045077770948410034,
-0.007498197257518768,
0.03312445804476738,
-0.09405501931905746,
0.007999680936336517,
0.00017547886818647385,
-0.05599250644445419,
-0.07181014120578766,
0.030648574233055115,
0.02147616073489189,
-0.01934962533414364,
0.08067324757575989,
-0.07865559309720993,
0.000011932510460610501,
-0.0723249539732933,
-0.12606371939182281,
0.021698759868741035,
-0.0672517716884613,
0.011537518352270126,
-0.12276905030012131,
-0.17878805100917816,
-0.005848981905728579,
0.05204681679606438,
-0.0257895365357399,
-0.04954242706298828,
-0.06130316108465195,
-0.08593835681676865,
0.019418971613049507,
-0.01918674446642399,
0.08458997309207916,
-0.07422292977571487,
0.0976468101143837,
0.03496040403842926,
0.05852235481142998,
-0.06102229282259941,
0.033584240823984146,
-0.10605329275131226,
0.03712766245007515,
-0.18268781900405884,
0.028077876195311546,
-0.037802647799253464,
0.07635728269815445,
-0.09719574451446533,
-0.07590635865926743,
-0.001005892874673009,
-0.00022226432338356972,
0.06155458465218544,
0.09921488165855408,
-0.17214272916316986,
-0.04863247275352478,
0.1747300773859024,
-0.08992677181959152,
-0.16386526823043823,
0.1397210657596588,
-0.05143260583281517,
0.03554574400186539,
0.070261150598526,
0.19845037162303925,
0.05776263773441315,
-0.0787862166762352,
0.011466209776699543,
0.0015727161662653089,
0.06100568547844887,
-0.047324348241090775,
0.08675281703472137,
-0.004099765792489052,
-0.014867344871163368,
0.005786554422229528,
-0.04723737761378288,
0.05523692071437836,
-0.0643303245306015,
-0.08191083371639252,
-0.04115846008062363,
-0.10133170336484909,
0.04032732918858528,
0.04563520476222038,
0.05584452673792839,
-0.12356887012720108,
-0.09118242561817169,
0.03956383094191551,
0.07092387974262238,
-0.0712403878569603,
0.027567630633711815,
-0.07209215313196182,
0.08032458275556564,
-0.060053929686546326,
-0.007378194481134415,
-0.15103460848331451,
-0.04919324070215225,
0.018023140728473663,
0.005755903664976358,
0.015173882246017456,
-0.010578298009932041,
0.06430632621049881,
0.07672931253910065,
-0.07026015967130661,
-0.036035437136888504,
0.000636062235571444,
0.014324475079774857,
-0.12205243855714798,
-0.20109641551971436,
-0.009471082128584385,
-0.025726597756147385,
0.15175898373126984,
-0.24352256953716278,
0.05353766307234764,
0.0039600529707968235,
0.07964219152927399,
0.0429433211684227,
0.00006918302096892148,
-0.02115294151008129,
0.06083539128303528,
-0.05319639667868614,
-0.06577267497777939,
0.0649372860789299,
0.016897426918148994,
-0.10519871115684509,
-0.014969756826758385,
-0.17371371388435364,
0.1873057633638382,
0.13534103333950043,
-0.07730191946029663,
-0.07210804522037506,
0.008583681657910347,
-0.03666000813245773,
-0.025408366695046425,
-0.039479345083236694,
-0.0017146096797659993,
0.12686923146247864,
-0.005848103668540716,
0.16489794850349426,
-0.1006791964173317,
-0.03769467771053314,
0.023997362703084946,
-0.0507160909473896,
0.008165555074810982,
0.1279526799917221,
0.07908961176872253,
-0.1477963924407959,
0.15097320079803467,
0.1880381852388382,
-0.07371245324611664,
0.14339855313301086,
-0.04858365282416344,
-0.04951607808470726,
-0.021714486181735992,
0.03129646182060242,
0.021271243691444397,
0.0998782142996788,
-0.10258927196264267,
0.005923358257859945,
0.00934518501162529,
0.021024145185947418,
0.011911006644368172,
-0.21219120919704437,
-0.02533806301653385,
0.04184502363204956,
-0.06076906621456146,
0.00010305452451575547,
-0.02238328382372856,
-0.007133892271667719,
0.10849206894636154,
0.0035298794973641634,
-0.08537206053733826,
0.04756096750497818,
-0.009033403359353542,
-0.09411256015300751,
0.21032682061195374,
-0.07944668084383011,
-0.17876452207565308,
-0.1381951868534088,
-0.04281413182616234,
-0.05825043469667435,
0.027154898270964622,
0.06718774884939194,
-0.0719393715262413,
-0.0389128178358078,
-0.1373918205499649,
-0.003841439029201865,
0.01792188175022602,
0.019148141145706177,
0.0023879138752818108,
0.007925175130367279,
0.083187997341156,
-0.0976121574640274,
-0.01304466463625431,
-0.03073570691049099,
-0.04490702226758003,
0.03104773908853531,
0.008936305530369282,
0.11188125610351562,
0.1512126326560974,
-0.029260000213980675,
-0.00116345367860049,
-0.040420107543468475,
0.21752674877643585,
-0.06227920949459076,
0.00184267433360219,
0.13996973633766174,
-0.0261502917855978,
0.052833426743745804,
0.14440445601940155,
0.04675893485546112,
-0.09681964665651321,
0.0385374054312706,
0.028767187148332596,
-0.03531377762556076,
-0.22263969480991364,
-0.030387207865715027,
-0.04735105112195015,
0.02693251520395279,
0.08875905722379684,
0.0388152189552784,
0.039504725486040115,
0.05455215647816658,
0.01295423973351717,
0.0739278718829155,
0.0024077678099274635,
0.07431045919656754,
0.10786841064691544,
0.04765675216913223,
0.125301331281662,
-0.06241609528660774,
-0.053092461079359055,
0.04101153090596199,
0.0017434220062568784,
0.18103985488414764,
0.017781797796487808,
0.16146211326122284,
0.046046581119298935,
0.14917899668216705,
0.006893618032336235,
0.06594288349151611,
-0.011703438125550747,
-0.030541809275746346,
-0.01954483613371849,
-0.05396067351102829,
-0.0314774252474308,
0.03262569382786751,
-0.09223385900259018,
0.06044025346636772,
-0.11040416359901428,
0.01563716121017933,
0.06287722289562225,
0.24034957587718964,
0.05283235013484955,
-0.3287091851234436,
-0.08565250784158707,
0.039009373635053635,
-0.02784554474055767,
-0.030259400606155396,
0.03717845678329468,
0.13420353829860687,
-0.05770875886082649,
0.046102508902549744,
-0.06902351975440979,
0.08833525329828262,
-0.04762713983654976,
0.04795825108885765,
0.048538051545619965,
0.0702141746878624,
-0.019976841285824776,
0.07504279911518097,
-0.2954113781452179,
0.26770904660224915,
0.015457142144441605,
0.062322404235601425,
-0.0550055168569088,
0.0003489605151116848,
0.029451770707964897,
0.09394027292728424,
0.0784391537308693,
-0.0166626013815403,
-0.08020754158496857,
-0.1813812404870987,
-0.063893161714077,
0.029304441064596176,
0.09108131378889084,
-0.04348396882414818,
0.10492976009845734,
-0.04128746688365936,
0.00047261849977076054,
0.08511529117822647,
-0.0004708178748842329,
-0.07713281363248825,
-0.09475400298833847,
-0.014321701601147652,
0.05568849667906761,
-0.029507191851735115,
-0.08365318179130554,
-0.08658727258443832,
-0.11666698008775711,
0.14170637726783752,
-0.05901797115802765,
-0.04307034611701965,
-0.09630434960126877,
0.05057641491293907,
0.040960658341646194,
-0.08241885900497437,
0.034577541053295135,
-0.0014762846985831857,
0.08984885364770889,
0.02404199168086052,
-0.06869272142648697,
0.1321689337491989,
-0.06807491183280945,
-0.1807311326265335,
-0.06181542947888374,
0.12115489691495895,
-0.0009871663060039282,
0.043161243200302124,
-0.0033902302384376526,
0.008675787597894669,
-0.013873079791665077,
-0.07305297255516052,
0.019699491560459137,
-0.014716026373207569,
0.058939021080732346,
0.018660182133316994,
-0.05090539529919624,
-0.019457029178738594,
-0.06567403674125671,
-0.04050973057746887,
0.1734517514705658,
0.2848614752292633,
-0.07501234859228134,
0.007621247787028551,
0.05488148331642151,
-0.06653420627117157,
-0.1956561803817749,
0.03322736918926239,
0.008520977571606636,
-0.004405906889587641,
0.047504447400569916,
-0.13944178819656372,
0.08615043759346008,
0.08420226722955704,
-0.028200512751936913,
0.10021708905696869,
-0.3077930510044098,
-0.1373213827610016,
0.12348775565624237,
0.1566198617219925,
0.11595666408538818,
-0.17483264207839966,
-0.042851898819208145,
-0.04999295994639397,
-0.09891506284475327,
0.12507855892181396,
-0.1508653312921524,
0.10528599470853806,
-0.0019117475021630526,
0.03854477405548096,
0.00704772025346756,
-0.050327200442552567,
0.12442796677350998,
-0.023919520899653435,
0.11309191584587097,
-0.07747527211904526,
-0.007413600105792284,
0.05657975748181343,
-0.06452164053916931,
0.03518204763531685,
-0.15077011287212372,
0.042831335216760635,
-0.07040069997310638,
-0.02648857794702053,
-0.04716392606496811,
0.031173886731266975,
-0.03892773017287254,
-0.0601109154522419,
-0.04157597944140434,
0.023738067597150803,
0.058042485266923904,
-0.0036070295609533787,
0.17429354786872864,
0.003214949509128928,
0.1583714634180069,
0.16088254749774933,
0.09214942157268524,
-0.04920509457588196,
-0.0014174390817061067,
-0.00043261409155093133,
-0.03690260276198387,
0.05493757873773575,
-0.16761687397956848,
0.03958910331130028,
0.11653127521276474,
-0.0005829146248288453,
0.14855331182479858,
0.07084634900093079,
-0.031533416360616684,
0.005057626869529486,
0.0665874257683754,
-0.17373809218406677,
-0.13214394450187683,
-0.021656805649399757,
-0.006679704412817955,
-0.11466953158378601,
0.0588495247066021,
0.13359034061431885,
-0.07682961225509644,
0.010954214259982109,
-0.005860788747668266,
0.026268556714057922,
-0.031935058534145355,
0.16111379861831665,
0.0662316307425499,
0.05450119823217392,
-0.07820159941911697,
0.09550175815820694,
0.0682736411690712,
-0.07494793832302094,
0.029115943238139153,
0.06611384451389313,
-0.09617415070533752,
-0.05197478458285332,
0.05620457977056503,
0.18699876964092255,
-0.02503998950123787,
-0.06748755276203156,
-0.15265682339668274,
-0.12692990899085999,
0.054142292588949203,
0.1913301944732666,
0.08548945933580399,
0.011377418413758278,
-0.025547701865434647,
-0.0012154433643445373,
-0.10974100977182388,
0.12681476771831512,
0.02973455935716629,
0.08156072348356247,
-0.15823739767074585,
0.1229080781340599,
-0.001416882500052452,
0.012534775771200657,
-0.021562485024333,
0.04538992419838905,
-0.11411844938993454,
0.0007319492870010436,
-0.1278410404920578,
0.009159402921795845,
-0.01693851314485073,
0.00757526233792305,
-0.0032991503830999136,
-0.044633835554122925,
-0.060068536549806595,
0.01152578555047512,
-0.09744665026664734,
-0.025370893999934196,
0.026583939790725708,
0.06091787666082382,
-0.12276574224233627,
-0.046008553355932236,
0.020915521308779716,
-0.0725385770201683,
0.0777648389339447,
0.0027579255402088165,
0.00647065881639719,
0.055925801396369934,
-0.16904747486114502,
0.04022311419248581,
0.057432591915130615,
0.015277039259672165,
0.03404838591814041,
-0.08811423927545547,
-0.015743868425488472,
0.0012849156046286225,
0.03727240115404129,
0.02504759654402733,
0.0920117199420929,
-0.12670497596263885,
0.009184671565890312,
-0.01733286865055561,
-0.05431423708796501,
-0.05140872299671173,
0.036092691123485565,
0.06863872706890106,
0.002336320700123906,
0.20257116854190826,
-0.09821265935897827,
0.009028272703289986,
-0.2050466686487198,
0.007111253682523966,
0.003927601035684347,
-0.12686263024806976,
-0.12275513261556625,
-0.050515975803136826,
0.05462636053562164,
-0.06714514642953873,
0.12490921467542648,
0.022332273423671722,
0.03429682180285454,
0.04222636669874191,
-0.026355935260653496,
0.04899090155959129,
0.03421563282608986,
0.23232987523078918,
0.01632382906973362,
-0.04895995929837227,
0.027293048799037933,
0.015141736716032028,
0.11309466511011124,
0.08116352558135986,
0.18652135133743286,
0.17412516474723816,
-0.04036867618560791,
0.10179304331541061,
0.03754846006631851,
-0.03875138610601425,
-0.12817469239234924,
0.054113682359457016,
-0.03831251338124275,
0.11837375164031982,
-0.01683025434613228,
0.20774097740650177,
0.10954300314188004,
-0.14944608509540558,
0.009137078188359737,
-0.06827592104673386,
-0.07811333984136581,
-0.11341341584920883,
-0.07548893243074417,
-0.09884049743413925,
-0.1380658745765686,
-0.01734725944697857,
-0.11681769043207169,
0.013684676960110664,
0.11698921769857407,
0.005611072294414043,
-0.023537741973996162,
0.16511523723602295,
0.01493478287011385,
0.01692316122353077,
0.06689044833183289,
-0.0006285369163379073,
-0.03464588150382042,
-0.07561397552490234,
-0.0910569578409195,
0.014494318515062332,
-0.022307883948087692,
0.021049395203590393,
-0.036735907196998596,
-0.02131788060069084,
0.04850684851408005,
-0.017865929752588272,
-0.10680463165044785,
0.015088842250406742,
0.02667883411049843,
0.04993436858057976,
0.05964036285877228,
0.0230523981153965,
-0.002332075498998165,
0.01052995678037405,
0.2370985746383667,
-0.0767831951379776,
-0.07403889298439026,
-0.08999582380056381,
0.2192516177892685,
0.020984571427106857,
0.0028598096687346697,
0.012107016518712044,
-0.09608354419469833,
0.02548580802977085,
0.21652360260486603,
0.1899687647819519,
-0.08795449137687683,
0.00041493645403534174,
-0.027665918692946434,
-0.0049947574734687805,
-0.024293463677167892,
0.10812633484601974,
0.11991823464632034,
0.01436555478721857,
-0.07609463483095169,
-0.029435070231556892,
-0.036629244685173035,
0.004299820400774479,
-0.047364454716444016,
0.07319722324609756,
0.03257916122674942,
0.009444116614758968,
-0.04049644619226456,
0.06978031247854233,
-0.05024905130267143,
-0.11329317092895508,
0.010798082686960697,
-0.20100289583206177,
-0.14480075240135193,
-0.02543501742184162,
0.11902495473623276,
-0.013162005692720413,
0.04355161264538765,
-0.03265614062547684,
0.008744900114834309,
0.04316992685198784,
-0.01976865530014038,
-0.06679729372262955,
-0.06259256601333618,
0.05803946033120155,
-0.12315073609352112,
0.21929052472114563,
-0.035683125257492065,
0.03653132542967796,
0.13411325216293335,
0.030985575169324875,
-0.0780690386891365,
0.10392243415117264,
0.04991234466433525,
-0.03498940169811249,
0.04393722489476204,
0.08101401478052139,
-0.036617524921894073,
0.12171266227960587,
0.06317857652902603,
-0.1242828369140625,
0.008268485777080059,
-0.04877341166138649,
-0.07616368681192398,
-0.05915394052863121,
-0.03983151540160179,
-0.06488032639026642,
0.13068613409996033,
0.1769043505191803,
-0.04858647659420967,
0.012919935397803783,
-0.04984667897224426,
0.006959294900298119,
0.06211143732070923,
0.021688342094421387,
-0.022379126399755478,
-0.2235584855079651,
0.019409241154789925,
0.06522654742002487,
0.005695655010640621,
-0.315968781709671,
-0.07225572317838669,
-0.020662177354097366,
-0.038236573338508606,
-0.10878732055425644,
0.09039802104234695,
0.13442082703113556,
0.045881252735853195,
-0.06789138168096542,
-0.07297933101654053,
-0.07235348224639893,
0.16695401072502136,
-0.11834219843149185,
-0.10034754872322083
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "meta-llama/Llama-2-7b-hf"} | null | RohanKumarMishra/llama_dp2 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-7b-hf",
"region:us"
] | 2024-02-06T05:57:41+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
41,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-meta-llama/Llama-2-7b-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.13237035274505615,
0.20393601059913635,
-0.002511046128347516,
0.02874687872827053,
0.07912357151508331,
0.019634027034044266,
0.04979075863957405,
0.13531364500522614,
0.020043307915329933,
0.10580451786518097,
0.0737132653594017,
0.11718367785215378,
0.10926163196563721,
0.20654499530792236,
0.003890186781063676,
-0.15440793335437775,
0.024214256554841995,
-0.08298544585704803,
0.006814117077738047,
0.1290476769208908,
0.14319083094596863,
-0.10468140989542007,
0.0831538662314415,
-0.014203370548784733,
0.0008161105797626078,
-0.03246506303548813,
-0.06674343347549438,
-0.015596466138958931,
0.04917285591363907,
0.02522817626595497,
0.05882670730352402,
-0.010089844465255737,
0.0929119735956192,
-0.26152917742729187,
0.018749000504612923,
0.04154228791594505,
0.009074261412024498,
0.08363344520330429,
0.0979103073477745,
-0.04074648395180702,
0.12078511714935303,
-0.024994686245918274,
0.13832204043865204,
0.09345067292451859,
-0.08226727694272995,
-0.233157217502594,
-0.06684722006320953,
0.07271547615528107,
0.18968668580055237,
0.08927863836288452,
-0.044125091284513474,
0.14097759127616882,
-0.07517150044441223,
0.02484818734228611,
0.04656748101115227,
-0.09290260076522827,
-0.06676048040390015,
0.0702265128493309,
0.13261590898036957,
0.0625041052699089,
-0.12113244831562042,
-0.03750992938876152,
0.03344248607754707,
0.044793009757995605,
0.06062353774905205,
0.005180627107620239,
0.16268815100193024,
0.034240271896123886,
-0.14592847228050232,
-0.05353321507573128,
0.14678435027599335,
0.01157673355191946,
-0.04636283218860626,
-0.21997328102588654,
-0.0027822081465274096,
-0.09489403665065765,
-0.022923149168491364,
-0.05228540673851967,
0.03324316069483757,
0.00603833794593811,
0.1196645051240921,
-0.042089227586984634,
-0.09635167568922043,
-0.029711460694670677,
0.0996040627360344,
0.05452839657664299,
0.02769845724105835,
-0.02099502831697464,
0.010653719305992126,
0.1290775090456009,
0.08296726644039154,
-0.1341402530670166,
-0.07021861523389816,
-0.0753326416015625,
-0.04316629841923714,
-0.03228989988565445,
0.03893959894776344,
0.019871119409799576,
0.07120058685541153,
0.2619621157646179,
-0.022196462377905846,
0.06401924788951874,
0.061033982783555984,
0.01709051802754402,
0.04062429443001747,
0.10795178264379501,
-0.03382651507854462,
-0.15705206990242004,
-0.007360270246863365,
0.10362072288990021,
-0.004135396331548691,
-0.02802850492298603,
-0.045986633747816086,
0.03152812272310257,
0.044165465980768204,
0.11501371115446091,
0.11203816533088684,
-0.019931387156248093,
-0.07717939466238022,
-0.05966082587838173,
0.19364216923713684,
-0.16149258613586426,
0.038572292774915695,
0.02467195875942707,
-0.006866174750030041,
-0.06484853476285934,
0.0073310090228915215,
0.016164373606443405,
-0.027354510501027107,
0.0603426918387413,
-0.0646006166934967,
-0.04179375246167183,
-0.1283673793077469,
-0.02387934736907482,
0.032629046589136124,
0.0170845165848732,
-0.0421639084815979,
-0.046661876142024994,
-0.08786044269800186,
-0.11000633984804153,
0.10926247388124466,
-0.05313732475042343,
-0.052913907915353775,
-0.02804330736398697,
-0.08941388875246048,
0.022293368354439735,
0.027490468695759773,
0.0755976140499115,
-0.02891632728278637,
0.052480049431324005,
0.003703000722452998,
0.059941843152046204,
0.0814133733510971,
0.027145687490701675,
-0.08097686618566513,
0.06685694307088852,
-0.19895170629024506,
0.07886288315057755,
-0.08557034283876419,
0.035526763647794724,
-0.16191443800926208,
-0.008882720954716206,
0.015485688112676144,
0.028551144525408745,
0.0418417863547802,
0.16628479957580566,
-0.21890771389007568,
-0.021091977134346962,
0.15901808440685272,
-0.10847076028585434,
-0.1374696046113968,
0.0436418242752552,
-0.04286689683794975,
0.18280568718910217,
0.028055870905518532,
0.010343263857066631,
0.09726855903863907,
-0.16840705275535583,
-0.02907063439488411,
-0.021288467571139336,
0.0036895605735480785,
0.07365763932466507,
0.09041544795036316,
-0.09089618921279907,
-0.0016403654590249062,
0.012144356034696102,
-0.06943254172801971,
-0.015110267326235771,
-0.04118245840072632,
-0.10628213733434677,
0.002018203027546406,
-0.09110194444656372,
0.023759065195918083,
0.0035124430432915688,
-0.09477277845144272,
-0.008542876690626144,
-0.1573835164308548,
-0.0652049109339714,
0.09409166127443314,
0.0002530320198275149,
-0.024702679365873337,
-0.10900412499904633,
0.06465248018503189,
-0.03883763402700424,
-0.026517964899539948,
-0.14125961065292358,
-0.023071611300110817,
0.01673055998980999,
-0.14134323596954346,
-0.01001854706555605,
-0.12183605134487152,
0.06567396223545074,
0.005137317348271608,
-0.0481104739010334,
-0.04708600044250488,
-0.004086394794285297,
0.0014921361580491066,
-0.05505292862653732,
-0.23444515466690063,
-0.028233496472239494,
-0.05085372179746628,
0.16539393365383148,
-0.2289838343858719,
0.044271692633628845,
0.014694449491798878,
0.11615854501724243,
-0.0018446118338033557,
-0.0661761611700058,
0.022094158455729485,
-0.07084274291992188,
-0.025033291429281235,
-0.07177132368087769,
-0.0071777342818677425,
0.00008959023398347199,
-0.029647991061210632,
0.015313859097659588,
-0.10952108353376389,
-0.053884293884038925,
0.100620798766613,
0.060472261160612106,
-0.14894865453243256,
0.008543584495782852,
-0.03779032453894615,
-0.06071627512574196,
-0.07427168637514114,
-0.0695083886384964,
0.0856412947177887,
0.052977994084358215,
0.03996400535106659,
-0.0812206119298935,
-0.07201940566301346,
0.005019875708967447,
-0.02742239646613598,
-0.005877636838704348,
0.11996077746152878,
0.07278608530759811,
-0.10015858709812164,
0.0890948474407196,
0.07567999511957169,
0.012905389070510864,
0.07863839715719223,
-0.028960783034563065,
-0.10615462064743042,
-0.03149069845676422,
0.05891314521431923,
0.0075002689845860004,
0.18196412920951843,
-0.07219336181879044,
0.05777830258011818,
0.046155888587236404,
-0.046635568141937256,
0.05089704319834709,
-0.09103982150554657,
0.0068960352800786495,
0.00045980032882653177,
-0.017081741243600845,
0.029599705711007118,
-0.020320137962698936,
0.006365274079144001,
0.07632698118686676,
0.05559656023979187,
0.02392573468387127,
0.023359429091215134,
-0.037590380758047104,
-0.1454712599515915,
0.18398217856884003,
-0.09283597022294998,
-0.235765740275383,
-0.15705986320972443,
0.0616452731192112,
0.049257904291152954,
-0.015689486637711525,
0.02697811834514141,
-0.055544715374708176,
-0.10059839487075806,
-0.08630408346652985,
-0.001965506933629513,
0.033574361354112625,
-0.05912783369421959,
-0.07473962754011154,
0.045523062348365784,
0.04523130878806114,
-0.11779510229825974,
0.02612960711121559,
0.06724361330270767,
-0.01014306303113699,
0.002122951438650489,
0.05421233922243118,
0.09625556319952011,
0.1871589571237564,
-0.0047584883868694305,
0.006493487861007452,
0.06463784724473953,
0.27302834391593933,
-0.16097134351730347,
0.10603976994752884,
0.1468280404806137,
-0.06509615480899811,
0.06928659975528717,
0.1811111718416214,
0.024897225201129913,
-0.0959320068359375,
0.024916043505072594,
0.02835996262729168,
-0.01960386149585247,
-0.2740720212459564,
-0.0512622706592083,
-0.015117009170353413,
-0.08622704446315765,
0.07128944247961044,
0.08718991279602051,
0.07891540229320526,
0.03938929736614227,
-0.05623466521501541,
-0.11011259257793427,
0.02521095983684063,
0.10682129859924316,
-0.01211885642260313,
0.003295447211712599,
0.08167944848537445,
-0.04613311216235161,
0.007927946746349335,
0.08699803054332733,
-0.01990879327058792,
0.1374768167734146,
0.04775961861014366,
0.09206060320138931,
0.08603846281766891,
0.10468525439500809,
-0.011216369457542896,
0.031460702419281006,
0.01713097095489502,
0.023083847016096115,
0.025577327236533165,
-0.0892123356461525,
0.00939508993178606,
0.11217135936021805,
0.02443520911037922,
0.02237142249941826,
0.016059260815382004,
-0.042084116488695145,
0.035355109721422195,
0.19778503477573395,
0.02863113395869732,
-0.21936152875423431,
-0.08315163850784302,
0.04950554668903351,
-0.07752750813961029,
-0.15846198797225952,
-0.0069001950323581696,
0.02585102617740631,
-0.16377925872802734,
0.015679948031902313,
-0.04114160314202309,
0.10047675669193268,
-0.07824478298425674,
-0.04026156663894653,
0.11029542237520218,
0.047400183975696564,
-0.01943347603082657,
0.05451195687055588,
-0.19536079466342926,
0.10843666642904282,
0.02992161363363266,
0.07536879926919937,
-0.08786998689174652,
0.09398660063743591,
0.006047630682587624,
-0.019160762429237366,
0.16931316256523132,
-0.0001144029592978768,
-0.049934081733226776,
-0.08560120314359665,
-0.09227954596281052,
0.0015766898868605494,
0.07818529009819031,
-0.12631447613239288,
0.0825691819190979,
-0.03569265082478523,
-0.024482207372784615,
-0.008127174340188503,
-0.08541606366634369,
-0.1325976550579071,
-0.14982733130455017,
0.05399367958307266,
-0.0976201519370079,
0.02554609440267086,
-0.08825770765542984,
-0.05347679927945137,
0.016768373548984528,
0.18224331736564636,
-0.21447692811489105,
-0.10864878445863724,
-0.14267513155937195,
-0.11213549226522446,
0.16079570353031158,
-0.042837124317884445,
0.08159231394529343,
0.00010400224709883332,
0.15704618394374847,
0.01110734511166811,
-0.015090357512235641,
0.08682332187891006,
-0.09437134861946106,
-0.19026298820972443,
-0.04887847229838371,
0.16311104595661163,
0.1444961428642273,
0.029530119150877,
-0.005065699107944965,
0.02549002133309841,
-0.06952440738677979,
-0.11216824501752853,
0.02609189972281456,
0.16361786425113678,
0.07300680130720139,
-0.012950204312801361,
-0.025871867313981056,
-0.0997539535164833,
-0.05963310971856117,
-0.04339827224612236,
-0.00898770522326231,
0.20425592362880707,
-0.06497634947299957,
0.14582973718643188,
0.10464579612016678,
-0.05606960505247116,
-0.21339629590511322,
0.03492094576358795,
0.04277806729078293,
0.026418045163154602,
0.04313372075557709,
-0.18166027963161469,
0.09741673618555069,
-0.014149999246001244,
-0.08650295436382294,
0.17498920857906342,
-0.17328102886676788,
-0.13439859449863434,
0.1159968227148056,
0.025544147938489914,
-0.21331895887851715,
-0.13972461223602295,
-0.10190334171056747,
-0.0198976993560791,
-0.126362144947052,
0.036111894994974136,
-0.0036879852414131165,
0.00850605871528387,
0.012948633171617985,
0.018173353746533394,
0.039593230932950974,
-0.05594787001609802,
0.21268853545188904,
-0.03937339782714844,
0.000047609177272534,
-0.050931964069604874,
-0.06770505011081696,
0.023772839456796646,
-0.0565045028924942,
0.12416863441467285,
-0.01210821233689785,
0.039195943623781204,
-0.17265570163726807,
-0.04285977780818939,
-0.058010976761579514,
0.03728554770350456,
-0.09242235124111176,
-0.0793662965297699,
-0.04483490809798241,
0.09155189245939255,
0.09041202813386917,
-0.018728721886873245,
0.0019666242878884077,
-0.09585212171077728,
0.07403325289487839,
0.20964933931827545,
0.20306745171546936,
0.0681707113981247,
-0.05247919633984566,
0.02836998738348484,
-0.03519117832183838,
0.04444263130426407,
-0.2148476094007492,
0.0430048331618309,
0.0631239265203476,
0.024400800466537476,
0.06267635524272919,
-0.01054441649466753,
-0.1590016484260559,
-0.07973737269639969,
0.08659059554338455,
-0.0608268640935421,
-0.16209019720554352,
-0.03262902423739433,
0.02129248157143593,
-0.2115628719329834,
-0.04105594381690025,
0.03599734604358673,
-0.014814808964729309,
-0.03840542584657669,
0.021407432854175568,
0.07970889657735825,
-0.028947602957487106,
0.1049608662724495,
0.09329938143491745,
0.09604475647211075,
-0.09774979948997498,
0.05453461781144142,
0.07179035246372223,
-0.031663764268159866,
0.03226640820503235,
0.1210775151848793,
-0.04315068572759628,
-0.046701591461896896,
0.08053972572088242,
0.11871292442083359,
-0.00035442441003397107,
-0.06335891038179398,
-0.0028557574842125177,
-0.0440225712954998,
0.054060470312833786,
0.10412941128015518,
0.036388467997312546,
0.0012024412862956524,
0.07687212526798248,
0.028011957183480263,
-0.09147296100854874,
0.12449978291988373,
0.06066809967160225,
0.02483541890978813,
-0.05523430183529854,
-0.038621995598077774,
-0.015819178894162178,
-0.0028008304070681334,
-0.01961326226592064,
-0.0014547118917107582,
-0.08309019356966019,
0.0061004795134067535,
-0.13227513432502747,
0.022323906421661377,
-0.07725922018289566,
0.00378548726439476,
0.036021001636981964,
-0.046576302498579025,
0.0013563713291659951,
-0.0008801636286079884,
-0.07430332899093628,
-0.05454954877495766,
-0.01629588007926941,
0.07790114730596542,
-0.13923588395118713,
0.03906119614839554,
0.07606222480535507,
-0.10726266354322433,
0.06878530234098434,
-0.007731399964541197,
0.008601504378020763,
0.0010856596054509282,
-0.13779860734939575,
0.05484551563858986,
-0.028775036334991455,
-0.006356567144393921,
0.005071246065199375,
-0.19585701823234558,
-0.00865773856639862,
-0.03182972967624664,
-0.0634872317314148,
0.019731810316443443,
-0.001073729363270104,
-0.11955288797616959,
0.1077868640422821,
0.004837313666939735,
-0.05712589994072914,
-0.0236744936555624,
0.042738161981105804,
0.0863419771194458,
-0.0053856209851801395,
0.12532570958137512,
-0.0293873380869627,
0.07612910121679306,
-0.17633569240570068,
-0.010070881806313992,
-0.015794692561030388,
0.05993741378188133,
-0.019834399223327637,
-0.03712667524814606,
0.06236843764781952,
-0.027145320549607277,
0.17265751957893372,
-0.004146610386669636,
0.07253459841012955,
0.0493277981877327,
0.008650471456348896,
0.04884583130478859,
0.07257263362407684,
0.06367837637662888,
-0.017801770940423012,
0.00016894470900297165,
0.04386947304010391,
-0.002970502246171236,
-0.051965516060590744,
-0.15762734413146973,
0.06277678161859512,
0.17842786014080048,
0.056998081505298615,
0.030175408348441124,
0.012138530611991882,
-0.12049488723278046,
-0.07329574972391129,
0.10845038294792175,
-0.021686408668756485,
-0.031095284968614578,
-0.06442723423242569,
0.21323516964912415,
0.1388614922761917,
-0.19825653731822968,
0.0702671930193901,
-0.06280558556318283,
-0.04658647999167442,
-0.14314492046833038,
-0.17366671562194824,
-0.059809304773807526,
-0.0547034814953804,
-0.026051264256238937,
-0.054700352251529694,
0.04570859298110008,
0.047346316277980804,
-0.0016739139100536704,
-0.02772514894604683,
0.1126171201467514,
0.02765420638024807,
-0.032165806740522385,
0.04451003298163414,
0.05619681254029274,
0.03682970255613327,
-0.09137814491987228,
0.007322985213249922,
0.0029695341363549232,
0.014342821203172207,
0.06777288764715195,
0.01613135077059269,
-0.06992621719837189,
0.02725713886320591,
-0.020467489957809448,
-0.12120343744754791,
0.042514219880104065,
-0.005491400603204966,
-0.02191038616001606,
0.14766326546669006,
0.039597559720277786,
0.008086306042969227,
-0.014769108034670353,
0.22978916764259338,
-0.079631008207798,
-0.08263124525547028,
-0.1393512636423111,
0.07894771546125412,
-0.07535439729690552,
0.020168637856841087,
0.02652786672115326,
-0.12502749264240265,
0.017455779016017914,
0.17437158524990082,
0.11967697739601135,
-0.01862110011279583,
0.005760727450251579,
0.04387581720948219,
0.003006097162142396,
-0.04732988774776459,
0.01692454144358635,
0.05290905013680458,
0.19558346271514893,
-0.0746847614645958,
0.054245725274086,
-0.01774757355451584,
-0.08059251308441162,
-0.020728278905153275,
0.09288354963064194,
-0.009933017194271088,
-0.004748775623738766,
-0.06074956804513931,
0.149005725979805,
-0.0759778842329979,
-0.20890262722969055,
0.06107410788536072,
-0.057474348694086075,
-0.13986754417419434,
-0.043588198721408844,
0.03270360454916954,
-0.02818191610276699,
-0.0004342520551290363,
0.05878293514251709,
-0.041880737990140915,
0.1787300854921341,
0.02775873802602291,
-0.04535049945116043,
-0.08805633336305618,
0.060195520520210266,
-0.15322564542293549,
0.28409940004348755,
0.02300625666975975,
0.06475372612476349,
0.11462150514125824,
-0.023716775700449944,
-0.14765876531600952,
0.016111766919493675,
0.11251717060804367,
-0.07146475464105606,
0.06923303008079529,
0.16616879403591156,
0.00888645276427269,
0.12871026992797852,
0.06517354398965836,
-0.04169101640582085,
0.03372213616967201,
-0.08477409183979034,
-0.04430316761136055,
-0.1301726996898651,
0.07585147768259048,
-0.09351208806037903,
0.15738072991371155,
0.11715016514062881,
-0.07169844210147858,
0.010452828370034695,
-0.02282477170228958,
0.09099912643432617,
0.012017005123198032,
0.10486294329166412,
0.01101954746991396,
-0.19380232691764832,
0.04388235881924629,
0.012521770782768726,
0.09230010956525803,
-0.21009819209575653,
-0.05027567222714424,
0.04558335989713669,
-0.022896859794855118,
-0.06855283677577972,
0.11809497326612473,
0.03357189893722534,
0.028112467378377914,
-0.037041857838630676,
-0.032784342765808105,
0.007307000923901796,
0.151776984333992,
-0.11639050394296646,
-0.019398227334022522
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "alexsherstinsky/Mistral-7B-v0.1-sharded"} | null | SudiptoPramanik/Mistral_RL_RL_ExtractiveSummary | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:alexsherstinsky/Mistral-7B-v0.1-sharded",
"region:us"
] | 2024-02-06T06:06:37+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-alexsherstinsky/Mistral-7B-v0.1-sharded #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-alexsherstinsky/Mistral-7B-v0.1-sharded #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
45,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-alexsherstinsky/Mistral-7B-v0.1-sharded #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.110109843313694,
0.20426276326179504,
-0.0033242744393646717,
0.02683502435684204,
0.07316219061613083,
0.012933526188135147,
0.07240378856658936,
0.13029123842716217,
0.0330861434340477,
0.12623506784439087,
0.0663594976067543,
0.11143319308757782,
0.11400832235813141,
0.21561922132968903,
-0.003767105983570218,
-0.17360882461071014,
0.020239561796188354,
-0.05931909754872322,
0.034331753849983215,
0.12480589002370834,
0.13765163719654083,
-0.09187468141317368,
0.07441461831331253,
-0.023067325353622437,
-0.007257964462041855,
-0.030011918395757675,
-0.06413144618272781,
-0.012694036588072777,
0.052279502153396606,
0.0378146693110466,
0.060410551726818085,
-0.007752938661724329,
0.08189915865659714,
-0.2682739794254303,
0.016325049102306366,
0.04730505868792534,
-0.011336163617670536,
0.08267117291688919,
0.10176274180412292,
-0.04810560867190361,
0.1253533810377121,
-0.037304703146219254,
0.1296939104795456,
0.08034759759902954,
-0.10833422094583511,
-0.22102825343608856,
-0.06735025346279144,
0.0868915244936943,
0.1795286238193512,
0.06535330414772034,
-0.04263977333903313,
0.12261921912431717,
-0.055373914539813995,
0.028244679793715477,
0.08744969964027405,
-0.11619655787944794,
-0.067912258207798,
0.06995866447687149,
0.13108506798744202,
0.08480945229530334,
-0.11932659894227982,
-0.03515271842479706,
0.033694859594106674,
0.0460745133459568,
0.06793062388896942,
0.007016418967396021,
0.157415509223938,
0.02952895313501358,
-0.1433580219745636,
-0.04947032406926155,
0.10342932492494583,
0.007777987979352474,
-0.04197976365685463,
-0.21659140288829803,
-0.010902638547122478,
-0.09021607786417007,
-0.03760097175836563,
-0.04981773719191551,
0.0327930748462677,
0.01047485787421465,
0.12073933333158493,
-0.05324931815266609,
-0.08308704942464828,
-0.009831502102315426,
0.11944828927516937,
0.06926008313894272,
0.010240095667541027,
-0.019501550123095512,
0.00468682823702693,
0.11847840994596481,
0.05970883369445801,
-0.12965521216392517,
-0.062441613525152206,
-0.0627671331167221,
-0.029873453080654144,
-0.024195287376642227,
0.04842666909098625,
0.02648993954062462,
0.042996060103178024,
0.27919477224349976,
-0.02499905787408352,
0.06445452570915222,
0.03833015263080597,
0.017506811767816544,
0.0209523793309927,
0.10952203720808029,
-0.027887746691703796,
-0.18805280327796936,
-0.00785139575600624,
0.1037183478474617,
0.00951753742992878,
-0.03065507300198078,
-0.05571940541267395,
0.02089713327586651,
0.04004978388547897,
0.12380597740411758,
0.10219241678714752,
-0.028415348380804062,
-0.0656670331954956,
-0.056997787207365036,
0.20572279393672943,
-0.15231499075889587,
0.057325515896081924,
0.03066612035036087,
0.000978308031335473,
-0.06201085448265076,
0.011775513179600239,
0.008082118816673756,
-0.03692951425909996,
0.09554214775562286,
-0.06441154330968857,
-0.038193073123693466,
-0.11660514026880264,
-0.04551363363862038,
0.0367552787065506,
-0.020857995375990868,
-0.04893016442656517,
-0.03855219483375549,
-0.08102472871541977,
-0.10572165250778198,
0.10145532339811325,
-0.054616548120975494,
-0.04785343259572983,
-0.03009726293385029,
-0.07338356971740723,
0.02454744465649128,
0.029810938984155655,
0.061975590884685516,
-0.026527537032961845,
0.044389721006155014,
-0.02305789664387703,
0.07123309373855591,
0.07824490964412689,
0.033632248640060425,
-0.08152387291193008,
0.06508684158325195,
-0.17742902040481567,
0.07929858565330505,
-0.062077075242996216,
0.033570464700460434,
-0.16534066200256348,
0.0007017551106400788,
0.004242916591465473,
0.03086933307349682,
0.05304974690079689,
0.15916115045547485,
-0.2037542313337326,
-0.0345250740647316,
0.17959627509117126,
-0.10244475305080414,
-0.11975224316120148,
0.03716188296675682,
-0.04131937026977539,
0.17497402429580688,
0.04257720708847046,
0.02343027852475643,
0.09054774045944214,
-0.15291258692741394,
-0.01978994719684124,
-0.031267520040273666,
0.014165452681481838,
0.0622839592397213,
0.07420126348733902,
-0.08111037313938141,
0.00047640083357691765,
0.005789510440081358,
-0.05705531686544418,
-0.021126434206962585,
-0.038276441395282745,
-0.09720899909734726,
0.008260915987193584,
-0.07907192409038544,
0.007491591852158308,
0.0029457921627908945,
-0.0937480553984642,
-0.01217607595026493,
-0.14669319987297058,
-0.025165997445583344,
0.07373563945293427,
0.003158573294058442,
-0.005840174853801727,
-0.08364780247211456,
0.04531695321202278,
-0.04651717469096184,
-0.012500773184001446,
-0.15108191967010498,
-0.0011525335721671581,
0.02214769273996353,
-0.14006948471069336,
0.012162993662059307,
-0.14084939658641815,
0.07386209070682526,
0.013708565384149551,
-0.05641097202897072,
-0.03984736651182175,
0.014070483855903149,
-0.014163503423333168,
-0.07220236212015152,
-0.22516661882400513,
-0.03213506564497948,
-0.05659179016947746,
0.1322016566991806,
-0.22785338759422302,
0.045285001397132874,
0.0014656432904303074,
0.11130665242671967,
0.012530438601970673,
-0.06164274737238884,
0.021894264966249466,
-0.056239765137434006,
-0.026936599984765053,
-0.071257583796978,
-0.003288415027782321,
0.0019245932344347239,
-0.030422238633036613,
0.02249748632311821,
-0.1347021609544754,
-0.07268781960010529,
0.087431900203228,
0.08374550938606262,
-0.14605119824409485,
0.005244195461273193,
-0.035594403743743896,
-0.057235270738601685,
-0.06911290436983109,
-0.07025576382875443,
0.07089311629533768,
0.051931776106357574,
0.049817245453596115,
-0.08891286700963974,
-0.07021832466125488,
-0.0014828727580606937,
-0.018785672262310982,
-0.023595018312335014,
0.12765425443649292,
0.07284100353717804,
-0.10459782183170319,
0.09470570087432861,
0.07270287722349167,
0.02697981894016266,
0.0994887426495552,
-0.007307672407478094,
-0.09977997839450836,
-0.03138745203614235,
0.05709034204483032,
0.02002185396850109,
0.15080347657203674,
-0.05995139852166176,
0.04183671623468399,
0.044179778546094894,
-0.04822225123643875,
0.036573950201272964,
-0.0953039899468422,
0.011296747252345085,
0.006154231261461973,
-0.01610223948955536,
0.02957276813685894,
-0.022515933960676193,
0.0028687939047813416,
0.09212246537208557,
0.06580419838428497,
0.02490052953362465,
0.010820328257977962,
-0.03876588121056557,
-0.1413283348083496,
0.17779025435447693,
-0.08620518445968628,
-0.2214423269033432,
-0.15582458674907684,
0.03303509205579758,
0.060906071215867996,
-0.010964687913656235,
0.03860332816839218,
-0.04540291801095009,
-0.08722398430109024,
-0.08975794911384583,
0.02394494228065014,
0.044791966676712036,
-0.05868641659617424,
-0.06814705580472946,
0.0333222895860672,
0.024159414693713188,
-0.13040386140346527,
0.023749617859721184,
0.04909688979387283,
-0.00034987536491826177,
-0.005188827868551016,
0.03241325169801712,
0.08447565883398056,
0.2065049111843109,
-0.0026565860025584698,
-0.00025402664323337376,
0.05427759140729904,
0.28095370531082153,
-0.15285564959049225,
0.12382353097200394,
0.12293418496847153,
-0.06467143446207047,
0.08293068408966064,
0.19206713140010834,
0.03356461226940155,
-0.08545693010091782,
0.015078777447342873,
0.035098034888505936,
-0.03581055626273155,
-0.26805275678634644,
-0.04256880655884743,
-0.024064278230071068,
-0.06810324639081955,
0.09134866297245026,
0.08128290623426437,
0.09407679736614227,
0.03345697373151779,
-0.0752536952495575,
-0.0801953673362732,
0.04511275887489319,
0.12162093818187714,
-0.050455618649721146,
0.017943501472473145,
0.08430361747741699,
-0.04937650263309479,
0.007476978003978729,
0.0860724225640297,
-0.01453783456236124,
0.13549737632274628,
0.057939957827329636,
0.11976924538612366,
0.07539068162441254,
0.062220171093940735,
0.0041070543229579926,
0.04859903082251549,
-0.009308382868766785,
0.026934761554002762,
0.014352924190461636,
-0.09536643326282501,
0.02176397480070591,
0.11496482044458389,
-0.000005564658295043046,
0.026965824887156487,
0.0200294591486454,
-0.07863914221525192,
0.039087630808353424,
0.2055603265762329,
0.035098589956760406,
-0.2077338844537735,
-0.08235063403844833,
0.06039128825068474,
-0.07191184908151627,
-0.15041004121303558,
-0.013588255271315575,
0.011162610724568367,
-0.15474948287010193,
0.019464297220110893,
-0.04525265842676163,
0.11251527070999146,
-0.06617746502161026,
-0.04283647984266281,
0.09621220827102661,
0.04985488951206207,
-0.04508543387055397,
0.03984001651406288,
-0.18906085193157196,
0.10983707010746002,
0.03138548135757446,
0.07320403307676315,
-0.08476196974515915,
0.08112826943397522,
-0.0009192607249133289,
-0.014862818643450737,
0.15639187395572662,
-0.0022264497820287943,
-0.07685430347919464,
-0.08429781347513199,
-0.06939982622861862,
-0.01644659787416458,
0.08421112596988678,
-0.1347922384738922,
0.07905580848455429,
-0.023771880194544792,
-0.03505420312285423,
-0.006907613482326269,
-0.09923306107521057,
-0.1062455028295517,
-0.16373415291309357,
0.05397550389170647,
-0.08048416674137115,
0.016202464699745178,
-0.07462850958108902,
-0.05205504596233368,
0.04472297057509422,
0.16162064671516418,
-0.20440195500850677,
-0.11403585225343704,
-0.14111113548278809,
-0.10442207753658295,
0.15391606092453003,
-0.05450672283768654,
0.08898589760065079,
-0.012385696172714233,
0.15804588794708252,
-0.01586785353720188,
-0.023965999484062195,
0.080719493329525,
-0.09016864746809006,
-0.1869654506444931,
-0.05040700361132622,
0.19184014201164246,
0.13546326756477356,
0.027443328872323036,
-0.012780127115547657,
0.029926499351859093,
-0.054875750094652176,
-0.1049741879105568,
0.02712760493159294,
0.13121086359024048,
0.06301523000001907,
-0.011507576331496239,
-0.03961675614118576,
-0.10761462152004242,
-0.06617847084999084,
-0.037552911788225174,
-0.013129501603543758,
0.21204674243927002,
-0.07131116837263107,
0.15669883787631989,
0.13428831100463867,
-0.06544041633605957,
-0.20483754575252533,
0.03753198683261871,
0.035385578870773315,
0.018236977979540825,
0.026067109778523445,
-0.19457732141017914,
0.07309531420469284,
-0.02141459286212921,
-0.07382556796073914,
0.17703941464424133,
-0.19694438576698303,
-0.13029493391513824,
0.09852635860443115,
0.01708853431046009,
-0.19515210390090942,
-0.1503404974937439,
-0.11023768782615662,
-0.01704263687133789,
-0.12067081779241562,
0.06257613003253937,
0.017869386821985245,
0.015143671073019505,
0.010425310581922531,
0.014392692595720291,
0.04404516518115997,
-0.04461923986673355,
0.19380077719688416,
-0.033769670873880386,
0.004635848104953766,
-0.05519195273518562,
-0.10341878980398178,
0.006317309103906155,
-0.06523780524730682,
0.11731242388486862,
-0.025350118055939674,
0.02443709224462509,
-0.15162475407123566,
-0.0441247783601284,
-0.06597283482551575,
0.02474210597574711,
-0.09377969801425934,
-0.08283840119838715,
-0.04770177975296974,
0.07903211563825607,
0.09041422605514526,
-0.017946941778063774,
0.037181757390499115,
-0.09022089838981628,
0.09729012101888657,
0.20138823986053467,
0.17175358533859253,
0.05355418846011162,
-0.045174892991781235,
0.027436474338173866,
-0.03736938163638115,
0.04741382226347923,
-0.223489910364151,
0.0372220054268837,
0.061571162194013596,
0.03466495871543884,
0.08293042331933975,
-0.003242778591811657,
-0.1632089763879776,
-0.08290805667638779,
0.08469267934560776,
-0.06166926026344299,
-0.16236701607704163,
-0.02590998262166977,
0.035343270748853683,
-0.20532868802547455,
-0.04493626952171326,
0.04636850208044052,
-0.019659148529171944,
-0.04349243640899658,
0.02372937835752964,
0.08205997943878174,
-0.02056795358657837,
0.09458784013986588,
0.08412887156009674,
0.09041876345872879,
-0.09277179092168808,
0.057113803923130035,
0.08248565346002579,
-0.018163925036787987,
0.019434349611401558,
0.14658372104167938,
-0.04009850695729256,
-0.037152811884880066,
0.07934527099132538,
0.11620350182056427,
-0.009561249054968357,
-0.04248342663049698,
0.01543293334543705,
-0.052525997161865234,
0.07117807865142822,
0.13478432595729828,
0.02001088857650757,
-0.010263524018228054,
0.0705440491437912,
0.02907724305987358,
-0.09167683124542236,
0.12405750900506973,
0.056559495627880096,
0.02387828752398491,
-0.02098764106631279,
-0.02241300418972969,
-0.017281534150242805,
-0.007748601958155632,
-0.01214271318167448,
-0.0017991504864767194,
-0.09985364228487015,
-0.00010818249575095251,
-0.11642701923847198,
0.023298067972064018,
-0.07721348851919174,
0.0023173552472144365,
0.015358264558017254,
-0.043099671602249146,
-0.0026699115987867117,
-0.008397807367146015,
-0.0753541886806488,
-0.0537470318377018,
-0.03262227028608322,
0.07683586329221725,
-0.14318186044692993,
0.02770192362368107,
0.07104771584272385,
-0.10877057909965515,
0.062038786709308624,
-0.007841352373361588,
0.015183927491307259,
0.005674487445503473,
-0.15054509043693542,
0.057142749428749084,
-0.024486707523465157,
-0.01725645363330841,
0.006643626838922501,
-0.1708439141511917,
-0.006305950228124857,
-0.051809463649988174,
-0.0719214603304863,
0.009533489122986794,
-0.015077232383191586,
-0.12315729260444641,
0.12094684690237045,
0.0017362898215651512,
-0.06565473973751068,
-0.014981938526034355,
0.05908970907330513,
0.07032697647809982,
-0.01675652526319027,
0.09807620942592621,
-0.028157642111182213,
0.08416876941919327,
-0.18083736300468445,
-0.0069732884876430035,
-0.014492437243461609,
0.03272204473614693,
-0.027742622420191765,
-0.041203975677490234,
0.05140266567468643,
-0.012445840053260326,
0.14870640635490417,
-0.003039774252101779,
0.07148155570030212,
0.04709630087018013,
0.006209264509379864,
0.03288114070892334,
0.07110495865345001,
0.060228172689676285,
-0.023647742345929146,
-0.014423592947423458,
0.026967685669660568,
0.0021891314536333084,
-0.04349159449338913,
-0.11992581933736801,
0.06914249807596207,
0.18553002178668976,
0.07927486300468445,
0.03253671154379845,
-0.0010351829696446657,
-0.12652595341205597,
-0.08451662212610245,
0.0840892568230629,
-0.006650243885815144,
-0.03133326768875122,
-0.06817347556352615,
0.22585780918598175,
0.14429612457752228,
-0.19503988325595856,
0.07997902482748032,
-0.04266722500324249,
-0.03197807818651199,
-0.13879773020744324,
-0.16527847945690155,
-0.05710893124341965,
-0.029848087579011917,
-0.035192620009183884,
-0.06382724642753601,
0.060045745223760605,
0.03486083075404167,
-0.0010674390941858292,
-0.009919303469359875,
0.10035693645477295,
0.019814493134617805,
-0.03784974664449692,
0.05296546593308449,
0.06662454456090927,
0.047949232161045074,
-0.08730470389127731,
0.012993368320167065,
0.0028715517837554216,
0.0020882494281977415,
0.062306344509124756,
0.029509833082556725,
-0.0573907233774662,
0.02843424677848816,
-0.016508016735315323,
-0.12342271208763123,
0.04832269996404648,
-0.005592701956629753,
-0.013650083914399147,
0.14911004900932312,
0.03406934067606926,
0.0032252620439976454,
-0.009044532664120197,
0.23616735637187958,
-0.06634797155857086,
-0.07783475518226624,
-0.12452603876590729,
0.08175703883171082,
-0.05504899099469185,
0.03096017800271511,
0.012054885737597942,
-0.12354312837123871,
0.014887244440615177,
0.1677706241607666,
0.12264934927225113,
-0.0039043996948748827,
0.009955290704965591,
0.040559761226177216,
0.010925398208200932,
-0.018291443586349487,
0.017071884125471115,
0.042369335889816284,
0.20928987860679626,
-0.07440926879644394,
0.07626623660326004,
-0.010689882561564445,
-0.06825059652328491,
-0.024058185517787933,
0.12240655720233917,
-0.013858120888471603,
-0.010844264179468155,
-0.061129167675971985,
0.1378892958164215,
-0.06585115939378738,
-0.21978257596492767,
0.05639028176665306,
-0.09162507206201553,
-0.13244394958019257,
-0.0419410765171051,
0.012980849482119083,
-0.02495395764708519,
0.017732052132487297,
0.06565281748771667,
-0.05679652467370033,
0.162874236702919,
0.027744030579924583,
-0.05252060294151306,
-0.10697993636131287,
0.05384554713964462,
-0.1422901302576065,
0.28400719165802,
0.021953271701931953,
0.03439553827047348,
0.10937730967998505,
-0.02023676596581936,
-0.1391659528017044,
0.010313108563423157,
0.10612982511520386,
-0.056747034192085266,
0.057422272861003876,
0.15852361917495728,
-0.007096059154719114,
0.12099551409482956,
0.0602305643260479,
-0.0642608031630516,
0.0335778146982193,
-0.05697423592209816,
-0.05715372413396835,
-0.12165220826864243,
0.07334904372692108,
-0.0846068412065506,
0.1480477899312973,
0.1273326873779297,
-0.07294914871454239,
-0.007207845337688923,
-0.01663087122142315,
0.07814791053533554,
0.021788625046610832,
0.12389518320560455,
0.013396648690104485,
-0.18079085648059845,
0.04512875899672508,
0.0093348678201437,
0.11271371692419052,
-0.22138561308383942,
-0.05580883100628853,
0.0416252501308918,
-0.019063347950577736,
-0.09296101331710815,
0.12022978067398071,
0.04253165423870087,
0.016048051416873932,
-0.02939530834555626,
-0.09239242970943451,
0.020172851160168648,
0.15283069014549255,
-0.09778184443712234,
-0.01708502136170864
] |
null | null | transformers | # miqu-1-120b

* EXL2: [2.4bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2) | 2.65bpw | [3.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2) | [4.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2) | [5.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2)
* GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/) | [IQ3_XXS](https://huggingface.co/wolfram/miqu-1-120b-GGUF)
* HF FP16: [wolfram/miqu-1-120b](https://huggingface.co/wolfram/miqu-1-120b)
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
Inspired by [Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2), [MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [goliath-120b](https://huggingface.co/alpindale/goliath-120b).
Thanks for the support, [CopilotKit](https://github.com/CopilotKit/CopilotKit) - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, [Lone Striker](https://huggingface.co/LoneStriker)!
## Prompt template: Mistral
```
<s>[INST] {prompt} [/INST]
```
See also: [🐺🐦⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates : LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/18ljvxb/llm_prompt_format_comparisontest_mixtral_8x7b/)
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
dtype: float16
merge_method: passthrough
slices:
- sources:
- layer_range: [0, 20]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [10, 30]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [20, 40]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [30, 50]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [40, 60]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [50, 70]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [60, 80]
model: 152334H/miqu-1-70b-sf
```
## Credits & Special Thanks
* original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
* leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
* f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
* mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
* mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
### Support
* [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| {"language": ["en", "de", "fr", "es", "it"], "library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["152334H/miqu-1-70b-sf"]} | text-generation | LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"en",
"de",
"fr",
"es",
"it",
"base_model:152334H/miqu-1-70b-sf",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:09:50+00:00 | [] | [
"en",
"de",
"fr",
"es",
"it"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # miqu-1-120b
!image/jpeg
* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw
* GGUF: Q2_K-Q5_K_M | IQ3_XXS
* HF FP16: wolfram/miqu-1-120b
This is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.
Inspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.
Thanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, Lone Striker!
## Prompt template: Mistral
See also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* 152334H/miqu-1-70b-sf
### Configuration
The following YAML configuration was used to produce this model:
## Credits & Special Thanks
* original (unreleased) model: mistralai (Mistral AI_)
* leaked model: miqudev/miqu-1-70b
* f16 model: 152334H/miqu-1-70b-sf
* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.
* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2
### Support
* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| [
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
85,
206,
44,
31,
4,
17,
28,
17,
107,
69,
43
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140## Merge Details### Merge Method\n\nThis model was merged using the passthrough merge method.### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.08452296257019043,
0.07996358722448349,
-0.006216804962605238,
0.07489742338657379,
0.062447689473629,
0.048032015562057495,
0.10850238054990768,
0.11730548739433289,
0.04997159540653229,
0.1304655373096466,
0.019059622660279274,
0.0882313996553421,
0.10314993560314178,
0.1546645164489746,
-0.0008784975507296622,
-0.13700617849826813,
0.03406155854463577,
-0.032812073826789856,
-0.05265757441520691,
0.07489487528800964,
0.07002948969602585,
-0.049538981169462204,
0.07827795296907425,
0.01992727816104889,
-0.0794241651892662,
-0.02351442351937294,
-0.03350690379738808,
-0.0034194772597402334,
0.06382977962493896,
0.09272785484790802,
0.010276054963469505,
0.01770416460931301,
0.010751184076070786,
-0.1683615893125534,
0.032938648015260696,
0.03592948243021965,
-0.017932593822479248,
0.07718798518180847,
0.10788709670305252,
-0.018908003345131874,
0.021088771522045135,
-0.10027661174535751,
-0.02568136155605316,
0.08249933272600174,
-0.10719147324562073,
-0.1474289745092392,
-0.18762712180614471,
0.15491966903209686,
0.042507048696279526,
0.03724713996052742,
-0.011902958154678345,
0.1333865076303482,
0.050249699503183365,
0.06016452983021736,
0.24593202769756317,
-0.22131793200969696,
-0.061919860541820526,
0.028957664966583252,
0.05541115999221802,
-0.04985284432768822,
-0.021356018260121346,
0.01992829702794552,
0.04054830223321915,
0.02191905304789543,
-0.020568734034895897,
-0.04234932363033295,
0.13632027804851532,
-0.05843184515833855,
-0.13171182572841644,
-0.02820662222802639,
0.09477535635232925,
0.04355064406991005,
-0.060202740132808685,
-0.1276080310344696,
-0.0682656466960907,
-0.03270931541919708,
-0.0022935839369893074,
-0.04131745919585228,
-0.002290143631398678,
-0.011816466227173805,
0.08523242175579071,
-0.08333887904882431,
-0.05337991565465927,
-0.020235886797308922,
-0.06682160496711731,
0.16135647892951965,
0.018354052677750587,
0.013679140247404575,
-0.013255320489406586,
0.07744071632623672,
-0.1265559196472168,
-0.1336318850517273,
-0.05428304150700569,
-0.03462158143520355,
-0.05692019686102867,
-0.05218517407774925,
0.01661340333521366,
-0.07750479131937027,
0.08107020705938339,
0.175857275724411,
-0.11741548776626587,
0.06220667064189911,
-0.009697522968053818,
0.030903007835149765,
0.024629943072795868,
0.08424429595470428,
-0.08733733743429184,
-0.08122188597917557,
0.04335138201713562,
0.05078619346022606,
0.05994807183742523,
0.0019238290842622519,
0.00010135882621398196,
-0.02575274370610714,
0.0019080113852396607,
0.07156599313020706,
0.06952427327632904,
0.029285242781043053,
-0.07679720222949982,
-0.05144278332591057,
0.17693321406841278,
-0.10877041518688202,
0.05875063315033913,
0.038095518946647644,
-0.010931058786809444,
-0.015534471720457077,
0.026578398421406746,
-0.028613878414034843,
-0.05422482639551163,
0.014574683271348476,
-0.052837468683719635,
0.006087095942348242,
-0.04121243208646774,
-0.04512706398963928,
0.052960656583309174,
0.044456131756305695,
-0.05848181992769241,
-0.12249575555324554,
-0.08548451960086823,
-0.056048691272735596,
0.033982738852500916,
-0.07165446132421494,
0.009947382844984531,
0.03666377440094948,
-0.08962082862854004,
0.029198521748185158,
0.02603435143828392,
0.022042686119675636,
-0.022196292877197266,
0.013299040496349335,
0.01655365526676178,
0.02530047670006752,
-0.031476203352212906,
-0.009267413057386875,
-0.05801313742995262,
0.08978775143623352,
-0.19187140464782715,
0.07319818437099457,
-0.06753759831190109,
-0.0458916574716568,
-0.1367833912372589,
-0.022929007187485695,
0.0001557070208946243,
-0.021685365587472916,
0.08853033930063248,
0.14028236269950867,
-0.15254990756511688,
-0.01800709031522274,
0.133183553814888,
-0.09939759224653244,
-0.08947071433067322,
0.09990517795085907,
0.010765177197754383,
-0.005259351339191198,
0.010271748527884483,
0.08585681021213531,
0.12536770105361938,
-0.056073881685733795,
-0.0745566338300705,
-0.10543492436408997,
-0.02353603020310402,
0.10377749055624008,
0.036487940698862076,
-0.06648851186037064,
0.006608173716813326,
0.0013413212727755308,
-0.03926616534590721,
-0.02470051497220993,
-0.01713505946099758,
-0.04516416788101196,
-0.051653604954481125,
-0.03216031193733215,
0.019755378365516663,
-0.0311356820166111,
-0.04306983947753906,
-0.06874691694974899,
-0.0662546455860138,
-0.03937284275889397,
0.10621660202741623,
0.0002958594122901559,
0.013465981930494308,
-0.06993132084608078,
0.12132570147514343,
-0.043282829225063324,
0.019509179517626762,
-0.13229045271873474,
-0.062326252460479736,
0.05427157133817673,
-0.17642687261104584,
0.057915039360523224,
-0.07802051305770874,
0.07007356733083725,
0.08792263269424438,
-0.03142979368567467,
-0.04443249851465225,
0.014312747865915298,
-0.011503690853714943,
-0.05503445118665695,
-0.13744132220745087,
-0.05339666083455086,
-0.03718583658337593,
0.1360798180103302,
-0.07717297226190567,
0.03823299705982208,
0.04849797114729881,
0.17298123240470886,
0.013948954641819,
-0.06230805069208145,
0.059276606887578964,
0.013416923582553864,
-0.01901276968419552,
-0.04254152253270149,
0.022224929183721542,
-0.015195275656878948,
-0.06335578858852386,
0.06797880679368973,
-0.16682805120944977,
-0.09226492792367935,
0.08205856382846832,
0.11543388664722443,
-0.10825490206480026,
-0.0011853284668177366,
-0.0016989412251859903,
-0.03144387900829315,
-0.02349192649126053,
-0.07283974438905716,
0.139337420463562,
0.03313908725976944,
0.05388158559799194,
-0.05538925528526306,
-0.07884406298398972,
0.012225146405398846,
-0.024691954255104065,
-0.07743193954229355,
0.12690365314483643,
0.02291349321603775,
-0.19036413729190826,
0.1086549460887909,
0.10671419650316238,
0.024346476420760155,
0.12972186505794525,
0.010163333266973495,
-0.020192109048366547,
-0.11323220282793045,
0.07214321941137314,
0.06328365206718445,
0.08764441311359406,
0.009260048158466816,
0.06931837648153305,
0.03807743266224861,
0.00207220995798707,
0.013925828970968723,
-0.11039725691080093,
0.03326145187020302,
0.019972439855337143,
-0.035299137234687805,
0.07873526960611343,
0.04182936251163483,
0.027880189940333366,
0.08705969154834747,
0.017179058864712715,
0.07423700392246246,
-0.01186748594045639,
-0.03507009521126747,
-0.09978814423084259,
0.13680647313594818,
-0.10421305149793625,
-0.1235826313495636,
-0.15689323842525482,
0.023495908826589584,
-0.1045532375574112,
-0.031237754970788956,
0.04314960539340973,
-0.04964744672179222,
-0.04190392047166824,
-0.09863945841789246,
0.07575149089097977,
0.04664899781346321,
-0.031183557584881783,
0.029061678797006607,
-0.04268911853432655,
0.05956493318080902,
-0.1079893708229065,
-0.03650837019085884,
0.0311584435403347,
0.003986986353993416,
0.03826166316866875,
0.06534901261329651,
0.09445244818925858,
0.06816873699426651,
0.010970757342875004,
-0.007111722603440285,
0.0249604694545269,
0.2451404184103012,
-0.09102492034435272,
0.08363116532564163,
0.1731153130531311,
0.012367261573672295,
0.08963974565267563,
0.1418784111738205,
0.0519220232963562,
-0.06685937941074371,
-0.013835576362907887,
0.023638667538762093,
-0.024145755916833878,
-0.19078323245048523,
-0.061981018632650375,
-0.049455784261226654,
0.02179974503815174,
0.016594912856817245,
0.056652046740055084,
-0.03246590495109558,
0.04957956075668335,
-0.06632956862449646,
-0.023145724087953568,
0.04272812604904175,
0.07885605841875076,
0.04942423850297928,
0.05232582241296768,
0.032079048454761505,
-0.030630147084593773,
-0.01874661073088646,
0.10800144821405411,
0.002024435205385089,
0.0828050896525383,
0.005763496737927198,
0.15540826320648193,
0.042188387364149094,
0.05763785168528557,
-0.006694222800433636,
0.05031910166144371,
0.012587900273501873,
0.01727849431335926,
0.011400274932384491,
-0.0922514945268631,
0.017045721411705017,
0.037224024534225464,
0.031022431328892708,
0.08605832606554031,
-0.019477766007184982,
-0.042653512209653854,
0.06828916072845459,
0.20636118948459625,
0.054764457046985626,
-0.18104644119739532,
-0.052608687430620193,
0.05323893949389458,
-0.005172972101718187,
-0.0441189780831337,
-0.0574423223733902,
0.0601525716483593,
-0.1421785205602646,
0.10297753661870956,
-0.02837306633591652,
0.08344317972660065,
-0.0768112912774086,
-0.025202374905347824,
-0.002551008015871048,
0.14314106106758118,
-0.007685830816626549,
0.009870217181742191,
-0.11473074555397034,
0.09677925705909729,
0.047673650085926056,
0.050363969057798386,
-0.0467650331556797,
0.05014784261584282,
0.06751372665166855,
-0.016121957451105118,
0.1009698212146759,
0.015580941922962666,
-0.08597638458013535,
-0.13121242821216583,
-0.07546716183423996,
-0.02328302152454853,
0.09724711626768112,
-0.07956860214471817,
0.11797893047332764,
-0.006677214056253433,
-0.041914381086826324,
-0.051285892724990845,
0.08619783818721771,
-0.16726025938987732,
-0.10657192021608353,
0.08049735426902771,
0.035933516919612885,
-0.00293326354585588,
-0.07889305055141449,
-0.03560107201337814,
-0.06792481988668442,
0.20780132710933685,
-0.12493043392896652,
-0.06458897888660431,
-0.12241725623607635,
-0.02887725457549095,
0.14838698506355286,
-0.09519194811582565,
0.016095273196697235,
-0.0620892271399498,
0.09675031155347824,
-0.05750473216176033,
-0.09322291612625122,
0.03633593022823334,
-0.0772782415151596,
-0.22163733839988708,
-0.029905345290899277,
0.1838245987892151,
0.031322285532951355,
0.03489852324128151,
0.0206821970641613,
0.04595774784684181,
0.021382097154855728,
-0.10996323078870773,
0.04105592891573906,
0.12817256152629852,
0.020900849252939224,
0.09039270132780075,
-0.045222893357276917,
-0.14004118740558624,
-0.09013182669878006,
-0.05677556246519089,
0.0997968390583992,
0.29861730337142944,
-0.05223260074853897,
0.030688846483826637,
0.04147891700267792,
-0.07043230533599854,
-0.15785279870033264,
-0.038915667682886124,
0.04657349735498428,
0.01753225363790989,
0.01858132891356945,
-0.07110119611024857,
0.0366257019340992,
0.10373353213071823,
-0.01873767003417015,
0.15667074918746948,
-0.263369619846344,
-0.12704867124557495,
-0.006682985462248325,
0.06700930744409561,
-0.01957916095852852,
-0.1452975571155548,
-0.11674544960260391,
-0.03796976059675217,
-0.1995268017053604,
0.09981389343738556,
-0.02467156946659088,
0.0965239554643631,
-0.023898916319012642,
0.015778150409460068,
0.03433018922805786,
-0.059150923043489456,
0.19815395772457123,
-0.039659541100263596,
0.04237253591418266,
-0.08535932749509811,
0.011777906678617,
0.026407601311802864,
-0.05703800171613693,
0.0950261577963829,
-0.035693515092134476,
0.017066331580281258,
-0.07961391657590866,
0.0030619590543210506,
-0.08434901386499405,
0.06988205015659332,
-0.053291499614715576,
-0.002292127348482609,
-0.06203543022274971,
0.07322024554014206,
0.02764309197664261,
-0.02415998838841915,
0.024202430620789528,
-0.021172653883695602,
0.10087262094020844,
0.2350863218307495,
0.09775305539369583,
0.05048970505595207,
-0.06141071021556854,
-0.00026774543221108615,
-0.04018312692642212,
0.02579716220498085,
-0.035933900624513626,
0.04462524875998497,
0.08787034451961517,
0.02455063909292221,
0.10782863199710846,
0.01810472458600998,
-0.1601009964942932,
0.018065448850393295,
0.06646080315113068,
-0.13265873491764069,
-0.2303391844034195,
0.003492377931252122,
0.04154038056731224,
-0.05837193876504898,
0.01806606538593769,
0.1862204670906067,
0.0009367137681692839,
-0.04963487386703491,
0.022825637832283974,
0.05416732281446457,
-0.03536570817232132,
0.15667866170406342,
0.00837104581296444,
0.04538920149207115,
-0.061773963272571564,
0.07580430060625076,
0.04042122885584831,
0.0016077554319053888,
0.018714869394898415,
0.17067234218120575,
-0.05190492421388626,
-0.07857772707939148,
0.04102415591478348,
0.03600002080202103,
-0.021241383627057076,
-0.021828800439834595,
-0.027304736897349358,
-0.10624261945486069,
0.013062734156847,
0.12654060125350952,
0.02681851200759411,
-0.018991822376847267,
0.08323623985052109,
-0.004140008706599474,
-0.028763674199581146,
0.07244459539651871,
0.05491362512111664,
0.08497325330972672,
-0.059819940477609634,
0.06284529715776443,
-0.02697847969830036,
0.03915136307477951,
-0.0026654857210814953,
0.019256316125392914,
-0.12592482566833496,
-0.05442740023136139,
-0.08961838483810425,
0.005896841175854206,
-0.11454344540834427,
-0.029031215235590935,
-0.0207052081823349,
-0.00936585757881403,
-0.02115669846534729,
0.03387326002120972,
-0.051367443054914474,
-0.11562348157167435,
-0.07470326870679855,
0.08342844247817993,
-0.14418748021125793,
-0.01641484536230564,
0.06919168680906296,
-0.07516539096832275,
0.09250383824110031,
0.01708120107650757,
0.0034909669775515795,
-0.009268661960959435,
-0.11126343160867691,
-0.02617836929857731,
-0.013749503530561924,
0.020847713574767113,
0.02727784588932991,
-0.20117603242397308,
-0.002813884522765875,
-0.04270019009709358,
-0.03368119150400162,
0.01241485308855772,
0.15425018966197968,
-0.09736622869968414,
-0.040856342762708664,
-0.03006616234779358,
-0.044825147837400436,
-0.06249745562672615,
0.018799860030412674,
0.10122360289096832,
0.009257921017706394,
0.08748460561037064,
-0.05421772226691246,
0.07078760117292404,
-0.17954121530056,
-0.01305952575057745,
0.022268623113632202,
-0.06352261453866959,
0.042893052101135254,
0.006506424397230148,
0.07432542741298676,
-0.024414660409092903,
0.033989157527685165,
-0.08212705701589584,
-0.018906600773334503,
0.026504697278141975,
-0.028520310297608376,
-0.021283207461237907,
0.017214683815836906,
0.020455483347177505,
-0.022489050403237343,
-0.024124853312969208,
-0.0313778817653656,
0.020932979881763458,
-0.019497908651828766,
-0.05568772181868553,
0.1307651400566101,
0.10116901993751526,
0.042394958436489105,
0.06033344939351082,
0.0699099600315094,
-0.06935380399227142,
-0.019917812198400497,
-0.015662092715501785,
-0.0501786470413208,
0.05921950191259384,
-0.0570748895406723,
0.11598431318998337,
0.0971740260720253,
-0.16128812730312347,
0.08817395567893982,
-0.03033575415611267,
-0.025268414989113808,
-0.0654015764594078,
-0.1747601330280304,
-0.06647072732448578,
-0.021373050287365913,
-0.007110625971108675,
-0.07304833084344864,
0.04277768358588219,
0.03683268278837204,
0.0017412106972187757,
-0.0061371480114758015,
0.12685923278331757,
-0.04393889009952545,
-0.058910902589559555,
0.035390980541706085,
0.037289973348379135,
-0.015837162733078003,
0.04472239688038826,
-0.006923268549144268,
0.022451741620898247,
0.012978330254554749,
0.05661128833889961,
0.08223818242549896,
-0.00805328506976366,
0.02584826387465,
-0.028271563351154327,
-0.10848145186901093,
-0.0006025367183610797,
0.005287060514092445,
0.01833254098892212,
0.08337931334972382,
0.05019523203372955,
0.016249405220150948,
-0.015350534580647945,
0.0810377299785614,
-0.02552863396704197,
-0.014691580086946487,
-0.1009829118847847,
0.15841826796531677,
-0.0559045635163784,
0.002254345454275608,
-0.0015080315060913563,
-0.11343378573656082,
-0.011597293429076672,
0.12920117378234863,
0.16337698698043823,
-0.08000685274600983,
0.03376281261444092,
0.017094949260354042,
0.00689410325139761,
0.004397694021463394,
0.049244899302721024,
0.059573933482170105,
0.10981784015893936,
-0.046494387090206146,
0.12055498361587524,
-0.0517595037817955,
-0.032632842659950256,
-0.07947799563407898,
0.09397511184215546,
-0.047008223831653595,
0.014640634879469872,
-0.017156826332211494,
0.07035721093416214,
-0.032913386821746826,
-0.16172361373901367,
0.03033449873328209,
-0.12312323600053787,
-0.138545423746109,
-0.00011208858632016927,
-0.0077992831356823444,
0.001197756384499371,
0.07809285819530487,
0.009989638812839985,
0.010767842642962933,
0.13262617588043213,
0.004294542595744133,
-0.07419104129076004,
-0.056784018874168396,
0.0005215644487179816,
-0.09722831845283508,
0.18984216451644897,
0.024254996329545975,
0.04178498685359955,
0.13091936707496643,
-0.033293675631284714,
-0.15407414734363556,
0.05585271492600441,
0.0700296014547348,
-0.09110260009765625,
0.022162333130836487,
0.14707012474536896,
-0.005258307326585054,
0.030615337193012238,
0.07387614995241165,
-0.053392838686704636,
0.019069518893957138,
0.0722138062119484,
0.019377747550606728,
-0.10239636152982712,
0.08977871388196945,
-0.10703950375318527,
0.13002605736255646,
0.18838851153850555,
-0.03489691764116287,
0.012865101918578148,
-0.03789073973894119,
0.04828639328479767,
0.038723692297935486,
0.09065895527601242,
-0.034608691930770874,
-0.19317959249019623,
0.06994378566741943,
-0.023887865245342255,
0.08475272357463837,
-0.1766246110200882,
-0.1239243745803833,
-0.029406068846583366,
-0.021863967180252075,
-0.013726417906582355,
0.0984341949224472,
0.09909792244434357,
0.0073255919851362705,
-0.023081263527274132,
-0.1434156447649002,
-0.02458539977669716,
0.10146303474903107,
-0.08631070703268051,
-0.04105154797434807
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-turkish-informal2formal_6.02.2024_1
This model is a fine-tuned version of [alpcansoydas/t5-turkish-informal2formal_30.01.2024_3](https://huggingface.co/alpcansoydas/t5-turkish-informal2formal_30.01.2024_3) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1459
- Rouge1: 0.8953
- Rouge2: 0.7840
- Rougel: 0.8937
- Rougelsum: 0.8939
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0003
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|
| 0.2779 | 1.0 | 818 | 0.1459 | 0.8953 | 0.7840 | 0.8937 | 0.8939 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "metrics": ["rouge"], "base_model": "alpcansoydas/t5-turkish-informal2formal_30.01.2024_3", "model-index": [{"name": "t5-turkish-informal2formal_6.02.2024_1", "results": []}]} | text2text-generation | alpcansoydas/t5-turkish-informal2formal_6.02.2024_1 | [
"transformers",
"pytorch",
"tensorboard",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:alpcansoydas/t5-turkish-informal2formal_30.01.2024_3",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:10:17+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #base_model-alpcansoydas/t5-turkish-informal2formal_30.01.2024_3 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| t5-turkish-informal2formal\_6.02.2024\_1
========================================
This model is a fine-tuned version of alpcansoydas/t5-turkish-informal2formal\_30.01.2024\_3 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1459
* Rouge1: 0.8953
* Rouge2: 0.7840
* Rougel: 0.8937
* Rougelsum: 0.8939
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0003
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #base_model-alpcansoydas/t5-turkish-informal2formal_30.01.2024_3 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
86,
97,
4,
33
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #base_model-alpcansoydas/t5-turkish-informal2formal_30.01.2024_3 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.10181097686290741,
0.061611875891685486,
-0.0037703930865973234,
0.09674203395843506,
0.11984594911336899,
0.00984271988272667,
0.19014200568199158,
0.13951700925827026,
-0.09205296635627747,
0.05088312551379204,
0.15390239655971527,
0.12694700062274933,
0.03817252814769745,
0.12204194813966751,
-0.03428639471530914,
-0.2597179114818573,
0.007839681580662727,
0.0508592389523983,
-0.031944382935762405,
0.14921794831752777,
0.09021863341331482,
-0.1034056693315506,
0.09323163330554962,
0.017636839300394058,
-0.18413065373897552,
0.012381303124129772,
-0.0025706104934215546,
-0.08671612292528152,
0.1307617425918579,
0.02234610542654991,
0.10153795033693314,
0.016970954835414886,
0.02345523238182068,
-0.1537517011165619,
0.006151149980723858,
0.025346649810671806,
-0.003024822333827615,
0.08461272716522217,
0.07274528592824936,
-0.040615569800138474,
0.14330846071243286,
-0.09610270708799362,
0.027765776962041855,
0.0014979412080720067,
-0.13792051374912262,
-0.21406465768814087,
-0.07445520162582397,
0.07889217883348465,
0.07341007143259048,
0.08926086872816086,
-0.01855684444308281,
0.12115513533353806,
-0.024411365389823914,
0.12668786942958832,
0.2597205638885498,
-0.305442214012146,
-0.0855955258011818,
-0.02417733706533909,
0.0003788777394220233,
0.12071789056062698,
-0.06022868677973747,
-0.0060365344397723675,
0.044962428510189056,
0.02752271667122841,
0.11893267184495926,
-0.03104010969400406,
-0.0029730680398643017,
-0.024122392758727074,
-0.11572970449924469,
-0.07939242571592331,
0.21718718111515045,
0.024961987510323524,
-0.03199991583824158,
-0.060506243258714676,
-0.06739877909421921,
-0.1589602530002594,
-0.021875625476241112,
-0.014932130463421345,
0.004612253047525883,
-0.04490954056382179,
-0.05562059208750725,
-0.028272854164242744,
-0.09882859140634537,
-0.05104279890656471,
-0.029713286086916924,
0.20926916599273682,
0.058019619435071945,
0.003558886470273137,
-0.04114990308880806,
0.10871702432632446,
0.011744625866413116,
-0.16697871685028076,
0.03489070013165474,
0.03595810383558273,
0.036289047449827194,
-0.020621147006750107,
-0.04307003319263458,
-0.13272802531719208,
-0.008118066936731339,
0.09570266306400299,
-0.04137318953871727,
0.07357168942689896,
0.019255023449659348,
0.03943655267357826,
-0.1162978857755661,
0.1597418338060379,
-0.0018711871234700084,
-0.01782800443470478,
-0.023237988352775574,
0.08314425498247147,
0.027160022407770157,
-0.034891266375780106,
-0.12045213580131531,
-0.011291875503957272,
0.13852731883525848,
0.02572285756468773,
-0.05802721530199051,
0.07138864696025848,
-0.02143438532948494,
-0.03954404592514038,
-0.016003897413611412,
-0.1149502843618393,
-0.0095971105620265,
-0.019413704052567482,
-0.07924073189496994,
-0.021745488047599792,
0.0007505894755013287,
-0.009907367639243603,
-0.05862530693411827,
0.07086965441703796,
-0.07166901230812073,
0.02564665675163269,
-0.04803839698433876,
-0.10865294933319092,
-0.008851365186274052,
-0.11468404531478882,
0.005185909569263458,
-0.11637642234563828,
-0.13970351219177246,
-0.000044530977902468294,
0.06078052148222923,
-0.034587107598781586,
-0.041961148381233215,
-0.061392929404973984,
-0.07313496619462967,
0.034080956131219864,
-0.006821910385042429,
0.06720937043428421,
-0.060387298464775085,
0.10830751806497574,
0.06709529459476471,
0.07331511378288269,
-0.007329541724175215,
0.0514528714120388,
-0.08919697254896164,
0.027737755328416824,
-0.15916985273361206,
0.07652647793292999,
-0.015221931040287018,
0.058796074241399765,
-0.11203745752573013,
-0.09730008244514465,
-0.014544610865414143,
-0.01669994555413723,
0.0930284634232521,
0.11914361268281937,
-0.15026292204856873,
-0.09737352281808853,
0.2136785089969635,
-0.07043146342039108,
-0.1344224065542221,
0.13574959337711334,
-0.04279310256242752,
0.08172143250703812,
0.07194773852825165,
0.2294415831565857,
0.07406400889158249,
-0.05180557072162628,
0.001464979606680572,
-0.02979561686515808,
0.08172641694545746,
-0.0023289499804377556,
0.08672679215669632,
-0.01391013152897358,
0.030286848545074463,
0.02546975202858448,
-0.046130768954753876,
0.05491939187049866,
-0.07480288296937943,
-0.09442110359668732,
-0.028775880113244057,
-0.07407807558774948,
0.0925518125295639,
0.03519904613494873,
0.11376979947090149,
-0.1285625398159027,
-0.10394085943698883,
0.017095984891057014,
0.0613294392824173,
-0.07643218338489532,
0.03268000856041908,
-0.07524973899126053,
0.0942225232720375,
-0.08783844113349915,
-0.003670081263408065,
-0.1672687977552414,
0.018950529396533966,
0.022957265377044678,
-0.00001301388238061918,
0.03909243643283844,
-0.019488409161567688,
0.0825209990143776,
0.06082300841808319,
-0.05639715865254402,
-0.045561421662569046,
-0.04740707576274872,
-0.012584353797137737,
-0.10949370265007019,
-0.1653580516576767,
-0.016793958842754364,
-0.012663265690207481,
0.11114251613616943,
-0.17091234028339386,
0.061538610607385635,
0.04424094781279564,
0.10887254774570465,
0.01873464696109295,
-0.02259748987853527,
-0.03344504535198212,
0.059462521225214005,
-0.037147052586078644,
-0.06045761704444885,
0.06693881005048752,
0.001774833188392222,
-0.12400529533624649,
-0.014151586219668388,
-0.16955730319023132,
0.17072972655296326,
0.11987755447626114,
-0.06862203776836395,
-0.08207407593727112,
0.0018271049484610558,
-0.05723249167203903,
-0.046632032841444016,
0.021453432738780975,
-0.02061007171869278,
0.2118164598941803,
0.0025423637125641108,
0.1644279658794403,
-0.10293269902467728,
-0.052792325615882874,
0.04056408628821373,
-0.050190553069114685,
-0.005947769153863192,
0.1300073266029358,
0.00003429173375479877,
-0.10066338628530502,
0.1348363161087036,
0.15902891755104065,
-0.05275651440024376,
0.1933184564113617,
-0.04050608351826668,
-0.08201316744089127,
-0.0425901859998703,
0.04122770205140114,
0.02102006785571575,
0.09012026339769363,
-0.17417556047439575,
-0.03932131826877594,
0.020568720996379852,
0.019645266234874725,
0.028477350249886513,
-0.19421613216400146,
-0.02073182724416256,
0.062142226845026016,
-0.052143096923828125,
0.005802306346595287,
0.014268919825553894,
-0.010503249242901802,
0.11947491019964218,
0.002716143848374486,
-0.042572781443595886,
0.041143424808979034,
0.004084636922925711,
-0.07517220824956894,
0.19486752152442932,
-0.06997594237327576,
-0.1846938282251358,
-0.1338246762752533,
-0.06445907801389694,
-0.09012523293495178,
0.011071081273257732,
0.0758904442191124,
-0.11028619110584259,
0.002581582171842456,
-0.0719212219119072,
0.04969855025410652,
-0.04247723147273064,
0.035199109464883804,
0.022101933136582375,
-0.029537271708250046,
0.052625373005867004,
-0.10498704016208649,
-0.023449748754501343,
-0.01096462644636631,
-0.03288283944129944,
0.06332077831029892,
-0.030475465580821037,
0.1033102497458458,
0.1235998198390007,
-0.03406396880745888,
0.03598373383283615,
-0.03412097319960594,
0.21431244909763336,
-0.08196060359477997,
-0.014663731679320335,
0.1150081530213356,
-0.02632780931890011,
0.07472608238458633,
0.10271891206502914,
0.048119161278009415,
-0.0794152170419693,
-0.007523802109062672,
0.009025040082633495,
-0.045257240533828735,
-0.22278279066085815,
-0.019592000171542168,
-0.06232732534408569,
0.010680867359042168,
0.10982251167297363,
0.025315357372164726,
0.07860637456178665,
0.08566860109567642,
0.001063041272573173,
0.09191631525754929,
-0.014157532714307308,
0.08610057085752487,
0.13409946858882904,
0.05818171054124832,
0.12275241315364838,
-0.05498968064785004,
-0.06150852516293526,
0.030879486352205276,
0.005324648693203926,
0.2320970594882965,
0.003984534647315741,
0.21469250321388245,
0.038455624133348465,
0.16339260339736938,
0.0021527381613850594,
0.07572077214717865,
0.0025719411205500364,
-0.016472503542900085,
0.000015993846318451688,
-0.06271092593669891,
-0.05506846308708191,
0.022994155064225197,
-0.06533046811819077,
0.05795002356171608,
-0.12429087609052658,
0.027356702834367752,
0.07262527197599411,
0.25259828567504883,
-0.006780795753002167,
-0.3456774652004242,
-0.12098990380764008,
-0.01184921246021986,
-0.042960479855537415,
-0.03149734437465668,
0.002501430921256542,
0.12695004045963287,
-0.09747979789972305,
0.051334358751773834,
-0.09608257561922073,
0.0841212049126625,
-0.04723664000630379,
0.04889756441116333,
0.04000800848007202,
0.04167268052697182,
-0.0011162268929183483,
0.07313111424446106,
-0.2867940366268158,
0.28499722480773926,
0.01710350066423416,
0.06215784326195717,
-0.08336357027292252,
0.005536873824894428,
0.034010447561740875,
0.022026633843779564,
0.07950863242149353,
-0.004158548079431057,
-0.043630201369524,
-0.16552765667438507,
-0.08157411217689514,
0.020305786281824112,
0.1296718865633011,
-0.08035658299922943,
0.10995086282491684,
-0.02160222828388214,
0.0351717546582222,
0.038424406200647354,
-0.011666592210531235,
-0.06156231462955475,
-0.11336497217416763,
0.03399274870753288,
0.01639672927558422,
-0.03520967438817024,
-0.06920909136533737,
-0.11244980245828629,
-0.07313858717679977,
0.15906526148319244,
-0.034436143934726715,
-0.08423308283090591,
-0.09818294644355774,
0.029023369774222374,
0.06851163506507874,
-0.1070205569267273,
0.00559099530801177,
-0.0063683767803013325,
0.06898518651723862,
0.0018762890249490738,
-0.05357709154486656,
0.08526494354009628,
-0.04157070070505142,
-0.15954048931598663,
-0.03084871731698513,
0.15940937399864197,
0.023719480261206627,
0.057805322110652924,
-0.012420417740941048,
0.0065482305362820625,
-0.023215878754854202,
-0.07542426139116287,
0.026205211877822876,
-0.03272281587123871,
0.03700881823897362,
-0.012929373420774937,
-0.012629536911845207,
0.02184862457215786,
-0.10916230827569962,
-0.031138738617300987,
0.17712602019309998,
0.26812535524368286,
-0.07631296664476395,
0.02100907638669014,
0.015626827254891396,
-0.06457614153623581,
-0.12672756612300873,
0.009819753468036652,
0.08756770938634872,
0.018674900755286217,
0.020750658586621284,
-0.19130343198776245,
0.03936072066426277,
0.08069387078285217,
0.009152167476713657,
0.0837542712688446,
-0.36614495515823364,
-0.13261638581752777,
0.0999104380607605,
0.14348582923412323,
0.09912162274122238,
-0.1622820496559143,
-0.04865929111838341,
0.0013189454330131412,
-0.12344609200954437,
0.06616194546222687,
-0.0748533308506012,
0.12166357785463333,
-0.040235165506601334,
0.11099393665790558,
0.022283677011728287,
-0.042201653122901917,
0.1486169844865799,
0.0077133942395448685,
0.06816872209310532,
-0.08273442834615707,
-0.03531038016080856,
0.12004191428422928,
-0.05789972469210625,
0.03018450364470482,
-0.1326378583908081,
0.029217949137091637,
-0.13022524118423462,
-0.01986698806285858,
-0.074228435754776,
0.03284383565187454,
-0.04794073477387428,
-0.060460932552814484,
-0.030699748545885086,
-0.010741764679551125,
0.0642431378364563,
7.285459560080199e-7,
0.14894706010818481,
0.01869790442287922,
0.1456519514322281,
0.11995179951190948,
0.08594946563243866,
-0.05464702844619751,
-0.04274451732635498,
-0.02194887213408947,
0.0005226879147812724,
0.05068981647491455,
-0.14119495451450348,
0.01026369258761406,
0.15304474532604218,
0.02027112804353237,
0.13429901003837585,
0.05428396537899971,
-0.04488854110240936,
0.03278379514813423,
0.06303965300321579,
-0.18197841942310333,
-0.12710262835025787,
-0.05378090217709541,
-0.0283834058791399,
-0.1129990965127945,
0.047297555953264236,
0.12627190351486206,
-0.07817046344280243,
-0.003470728173851967,
-0.0161071065813303,
-0.02556363120675087,
-0.04297275096178055,
0.17492397129535675,
0.01892680861055851,
0.051866285502910614,
-0.06963175535202026,
0.07801797986030579,
0.043230678886175156,
-0.08576415479183197,
0.043507106602191925,
0.0652025043964386,
-0.08478115499019623,
-0.018976286053657532,
0.023433350026607513,
0.16069161891937256,
-0.03525422140955925,
-0.039790403097867966,
-0.16145093739032745,
-0.10248561948537827,
0.05793215334415436,
0.12551158666610718,
0.08829870820045471,
0.0036202017217874527,
-0.0629476010799408,
0.0165016558021307,
-0.09600582718849182,
0.10641011595726013,
0.08393952995538712,
0.058318477123975754,
-0.13879673182964325,
0.17611275613307953,
0.004808598663657904,
0.02725350670516491,
-0.011355222202837467,
-0.010403338819742203,
-0.0772920623421669,
0.013729717582464218,
-0.13358162343502045,
-0.010403970256447792,
-0.04367022216320038,
-0.003037757473066449,
-0.018442317843437195,
-0.04751039296388626,
-0.06132066622376442,
0.008573903702199459,
-0.11000841110944748,
-0.02460934966802597,
0.020847303792834282,
0.05600692704319954,
-0.09396295249462128,
-0.016440315172076225,
0.02400175854563713,
-0.08369381725788116,
0.07824627310037613,
0.08176157623529434,
-0.012604222632944584,
0.05941082909703255,
-0.049684878438711166,
0.05657840892672539,
0.03167901933193207,
0.005156008992344141,
0.0312734954059124,
-0.07190201431512833,
-0.0114958006888628,
-0.0007386408979073167,
0.0364668145775795,
0.03645218908786774,
0.0958385244011879,
-0.11878468096256256,
0.003865111619234085,
-0.016907813027501106,
-0.05163144692778587,
-0.08563288301229477,
0.09009746462106705,
0.06777748465538025,
0.027316700667142868,
0.16669517755508423,
-0.0910782441496849,
0.04104183241724968,
-0.1997675746679306,
-0.003028491511940956,
0.01346003357321024,
-0.13629204034805298,
-0.05578360706567764,
-0.0697542056441307,
0.04814240336418152,
-0.08336419612169266,
0.10209279507398605,
-0.01219112053513527,
0.028596581891179085,
0.024229414761066437,
-0.04173604026436806,
0.007977337576448917,
-0.0005199803854338825,
0.22552356123924255,
0.015481659211218357,
-0.04719659313559532,
0.03199794143438339,
0.02435985393822193,
0.08469726890325546,
0.10992047935724258,
0.1740136295557022,
0.1568870097398758,
-0.023076200857758522,
0.09581813216209412,
0.027372965589165688,
-0.023685768246650696,
-0.1935795098543167,
0.026144538074731827,
-0.0022023317869752645,
0.1069657951593399,
-0.0006473358371295035,
0.2164977788925171,
0.16253462433815002,
-0.15834060311317444,
0.03469013050198555,
-0.014769568108022213,
-0.08736377209424973,
-0.06848375499248505,
-0.13032905757427216,
-0.059282585978507996,
-0.13158847391605377,
-0.002222547773271799,
-0.10263914614915848,
0.025139905512332916,
0.06709454208612442,
0.016050927340984344,
-0.027075912803411484,
0.22163815796375275,
0.010782923549413681,
-0.022617900744080544,
0.09636949747800827,
-0.015889113768935204,
-0.028294166550040245,
-0.09032408148050308,
-0.08811494708061218,
-0.013487477786839008,
-0.01199270412325859,
0.03780350834131241,
-0.021614743396639824,
-0.044334445148706436,
0.00658451858907938,
-0.04137888923287392,
-0.1074608713388443,
0.0115130003541708,
0.02152378112077713,
0.06271093338727951,
0.03352680802345276,
0.014654319733381271,
-0.03071022406220436,
-0.031755801290273666,
0.2051951140165329,
-0.06112009659409523,
-0.032483603805303574,
-0.08032069355249405,
0.1420566290616989,
0.030438225716352463,
-0.025267917662858963,
0.043610040098428726,
-0.06280704587697983,
0.013433021493256092,
0.1964513212442398,
0.19104599952697754,
-0.013466610573232174,
0.008559931069612503,
-0.009501803666353226,
0.008263625204563141,
-0.0026240593288093805,
0.09177307039499283,
0.08617088943719864,
0.025386391207575798,
-0.07062721252441406,
-0.010062345303595066,
-0.06869460642337799,
-0.0016560503281652927,
-0.04542212560772896,
0.08302762359380722,
0.04103393107652664,
-0.010155830532312393,
-0.04522979259490967,
0.06735385209321976,
-0.06197439134120941,
-0.07633712887763977,
0.010979602113366127,
-0.2056117057800293,
-0.1499691903591156,
-0.002415182301774621,
0.07416319847106934,
0.006158818025141954,
0.07242055982351303,
-0.004163400270044804,
0.0013834183337166905,
0.061463724821805954,
0.009154584258794785,
-0.0795147716999054,
-0.07189831137657166,
0.09581196308135986,
-0.15852752327919006,
0.170205220580101,
-0.0347273014485836,
0.056775011122226715,
0.1203412115573883,
0.05790549889206886,
-0.05314851179718971,
0.058068450540304184,
0.0703013464808464,
0.010246634483337402,
0.02097349613904953,
0.11606108397245407,
-0.015813268721103668,
0.0633811503648758,
0.06834574788808823,
-0.10013575851917267,
-0.00017436707275919616,
-0.03020811453461647,
-0.00821281410753727,
-0.04678787291049957,
-0.036955635994672775,
-0.04906735196709633,
0.1197093054652214,
0.23484468460083008,
-0.043361131101846695,
-0.014522447250783443,
-0.061447013169527054,
-0.0010178148513659835,
0.10079289972782135,
0.02916586771607399,
-0.05196761339902878,
-0.2599298655986786,
-0.034780103713274,
0.06396578252315521,
-0.004587958566844463,
-0.24288693070411682,
-0.11759282648563385,
0.0016471671406179667,
-0.05642262473702431,
-0.07823417335748672,
0.08973872661590576,
0.09197276085615158,
0.05022721737623215,
-0.0534941628575325,
-0.05122749134898186,
-0.07836820185184479,
0.17734065651893616,
-0.1513393521308899,
-0.06777207553386688
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper tiny ZH
This model is a fine-tuned version of [ZhihCheng/whisper-base-zh](https://huggingface.co/ZhihCheng/whisper-base-zh) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0002
- Wer: 0.0
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 25
- training_steps: 100
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:---:|
| 0.0006 | 0.67 | 50 | 0.0004 | 0.0 |
| 0.0004 | 1.33 | 100 | 0.0002 | 0.0 |
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.0.1+cu117
- Datasets 2.16.1
- Tokenizers 0.15.0 | {"language": ["zh"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["ZhihCheng/Motor_jargon_dataset"], "metrics": ["wer"], "base_model": "ZhihCheng/whisper-base-zh", "model-index": [{"name": "Whisper tiny ZH", "results": []}]} | automatic-speech-recognition | ZhihCheng/whisper-base-zh_motor | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"generated_from_trainer",
"zh",
"dataset:ZhihCheng/Motor_jargon_dataset",
"base_model:ZhihCheng/whisper-base-zh",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:10:48+00:00 | [] | [
"zh"
] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #zh #dataset-ZhihCheng/Motor_jargon_dataset #base_model-ZhihCheng/whisper-base-zh #license-apache-2.0 #endpoints_compatible #region-us
| Whisper tiny ZH
===============
This model is a fine-tuned version of ZhihCheng/whisper-base-zh on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0002
* Wer: 0.0
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 25
* training\_steps: 100
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* Pytorch 2.0.1+cu117
* Datasets 2.16.1
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 25\n* training\\_steps: 100\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.0.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #zh #dataset-ZhihCheng/Motor_jargon_dataset #base_model-ZhihCheng/whisper-base-zh #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 25\n* training\\_steps: 100\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.0.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
92,
130,
4,
38
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #zh #dataset-ZhihCheng/Motor_jargon_dataset #base_model-ZhihCheng/whisper-base-zh #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 25\n* training\\_steps: 100\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.0.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
-0.14480377733707428,
0.07465021312236786,
-0.0016415056306868792,
0.0380377396941185,
0.09544845670461655,
-0.01628842018544674,
0.11316607147455215,
0.14466039836406708,
-0.025968952104449272,
0.09839455038309097,
0.09496816247701645,
0.08098836988210678,
0.07017631828784943,
0.1283062845468521,
-0.03424442932009697,
-0.30266162753105164,
0.03698320686817169,
0.01966640166938305,
-0.11439575999975204,
0.10120481997728348,
0.11178853362798691,
-0.12474527209997177,
0.030537275597453117,
0.03338062763214111,
-0.0852532833814621,
0.002916354686021805,
-0.014499606564640999,
-0.07933149486780167,
0.11086470633745193,
0.021892590448260307,
0.07390648126602173,
0.05264494940638542,
0.08817792683839798,
-0.2536051273345947,
0.02011079341173172,
0.05203999578952789,
0.038425322622060776,
0.05940106138586998,
0.08594220876693726,
-0.02298733964562416,
0.09656020253896713,
-0.09493564069271088,
0.06230567395687103,
0.05558585003018379,
-0.08383068442344666,
-0.32723456621170044,
-0.08784256875514984,
0.05926261097192764,
0.10961730778217316,
0.06861758977174759,
-0.03157905116677284,
0.1012996956706047,
-0.04385947808623314,
0.08006439357995987,
0.22718361020088196,
-0.26502248644828796,
-0.08500171452760696,
-0.021378453820943832,
0.07975143194198608,
0.054942578077316284,
-0.14254873991012573,
-0.007232654839754105,
0.04011200740933418,
0.020714975893497467,
0.09827413409948349,
0.00896085798740387,
0.04602494090795517,
0.007224544882774353,
-0.1484130173921585,
-0.04222511127591133,
0.15787483751773834,
0.07905381172895432,
-0.05292932689189911,
-0.09901472181081772,
-0.019533896818757057,
-0.13631553947925568,
-0.05576016753911972,
0.007440485525876284,
0.002492043189704418,
-0.038617029786109924,
-0.1070919930934906,
-0.014889554120600224,
-0.09578001499176025,
-0.10165780782699585,
0.0567774698138237,
0.20035794377326965,
0.06292308866977692,
-0.03196229413151741,
0.0020855076145380735,
0.10117843747138977,
0.05684851482510567,
-0.1612924188375473,
-0.03363063186407089,
0.018607506528496742,
-0.08663729578256607,
-0.01749579980969429,
-0.03134351596236229,
-0.042615313082933426,
0.03947294130921364,
0.15984560549259186,
-0.0008637066930532455,
0.1062079668045044,
0.020413264632225037,
0.01942146010696888,
-0.10786911100149155,
0.17164340615272522,
-0.052810125052928925,
-0.011574598029255867,
-0.0112614119425416,
0.13764476776123047,
-0.009544224478304386,
-0.024950368329882622,
-0.04613551124930382,
0.010278982110321522,
0.14649638533592224,
0.04417004436254501,
-0.03125085309147835,
0.03218136727809906,
-0.04550141468644142,
-0.01948929764330387,
-0.05693725496530533,
-0.10357481241226196,
0.00873278547078371,
0.035661548376083374,
-0.08231351524591446,
-0.027805708348751068,
0.009871634654700756,
0.028727205470204353,
-0.012572243809700012,
0.06403231620788574,
-0.06659847497940063,
-0.02257269248366356,
-0.07957915961742401,
-0.07242085784673691,
0.02275024726986885,
-0.019092915579676628,
0.009155785664916039,
-0.09222417324781418,
-0.13299231231212616,
-0.04799039661884308,
0.05229830741882324,
-0.01745651662349701,
-0.047995518893003464,
-0.06977573037147522,
-0.07536935806274414,
0.05175218731164932,
-0.019496073946356773,
0.1314370334148407,
-0.04853767901659012,
0.1206558346748352,
0.03183940052986145,
0.029841724783182144,
0.03433414176106453,
0.050265755504369736,
-0.04129686951637268,
0.063524030148983,
-0.08582361042499542,
0.06411495059728622,
-0.11317954957485199,
0.06933098286390305,
-0.13970455527305603,
-0.10407818108797073,
-0.016852568835020065,
-0.002061765408143401,
0.0974297896027565,
0.09994693100452423,
-0.1695949137210846,
-0.06642076373100281,
0.1913817673921585,
-0.08650234341621399,
-0.1375553160905838,
0.13346119225025177,
-0.028604527935385704,
0.0367622971534729,
0.04844789206981659,
0.2018475979566574,
0.11650966852903366,
-0.11154086142778397,
-0.0010546103585511446,
-0.05606837198138237,
0.12211829423904419,
0.02848011814057827,
0.1120704859495163,
-0.033506084233522415,
-0.0006750580505467951,
0.006492073647677898,
-0.052092019468545914,
0.05124761909246445,
-0.08720150589942932,
-0.09278664737939835,
-0.02864251658320427,
-0.10204067826271057,
0.04136021062731743,
0.06076033413410187,
0.049338966608047485,
-0.10186686366796494,
-0.12221254408359528,
0.022331517189741135,
0.13535554707050323,
-0.11422685533761978,
0.013507538475096226,
-0.103912353515625,
0.06118285283446312,
-0.011294064112007618,
-0.0071987733244895935,
-0.13906951248645782,
-0.04156129062175751,
0.03675936907529831,
-0.06539153307676315,
0.015286870300769806,
-0.0439477264881134,
0.08013657480478287,
0.05446917936205864,
-0.036919984966516495,
-0.10486440360546112,
-0.07267899066209793,
0.01076619140803814,
-0.07239548861980438,
-0.19387178122997284,
-0.07985884696245193,
-0.03303806111216545,
0.18173043429851532,
-0.24952487647533417,
0.03441120684146881,
0.05422557517886162,
0.12055299431085587,
0.03200702369213104,
-0.035263895988464355,
-0.0001746789348544553,
0.041162680834531784,
-0.013024669140577316,
-0.06887444108724594,
0.03424825891852379,
0.019343502819538116,
-0.12983819842338562,
0.022321952506899834,
-0.14655926823616028,
0.09677638858556747,
0.08799909055233002,
0.0029598716646432877,
-0.04352241009473801,
-0.06041548401117325,
-0.061783984303474426,
-0.0680822879076004,
-0.01820407062768936,
-0.02968519553542137,
0.1590127944946289,
0.015188110060989857,
0.11714086681604385,
-0.09991471469402313,
-0.05283195897936821,
0.0244252011179924,
-0.023124761879444122,
-0.012341695837676525,
0.13349223136901855,
0.0022770222276449203,
-0.08713886141777039,
0.09337054193019867,
0.08061724156141281,
-0.08533018827438354,
0.1875545084476471,
-0.09043846279382706,
-0.09707829356193542,
-0.021059049293398857,
0.050000905990600586,
0.04273104667663574,
0.14606614410877228,
-0.12349186092615128,
-0.014956273138523102,
0.017635125666856766,
0.0074659246020019054,
0.03224363923072815,
-0.21341320872306824,
-0.005434450227767229,
0.03157447278499603,
-0.06706905364990234,
-0.005250465590506792,
0.0205700621008873,
-0.015943489968776703,
0.07629947364330292,
-0.007489117328077555,
-0.03825266286730766,
0.0044056111946702,
-0.0375308096408844,
-0.1044522374868393,
0.1955929696559906,
-0.09882300347089767,
-0.12657617032527924,
-0.13846272230148315,
-0.011010454967617989,
-0.024384308606386185,
-0.013559085316956043,
0.04369019344449043,
-0.08679123222827911,
-0.0261685810983181,
-0.063188336789608,
0.010652126744389534,
-0.026607971638441086,
0.028740402311086655,
0.0033990079537034035,
0.0030871867202222347,
0.08956465125083923,
-0.10550288110971451,
0.012184987775981426,
-0.024666013196110725,
-0.0497647225856781,
0.025453045964241028,
-0.0031822589226067066,
0.09352428466081619,
0.1355697065591812,
0.0317867249250412,
0.032897934317588806,
-0.03396943584084511,
0.14431923627853394,
-0.09306292235851288,
-0.01801316998898983,
0.1539735049009323,
-0.007611046079546213,
0.05272180587053299,
0.10299016535282135,
0.03673481196165085,
-0.08408026397228241,
0.034322045743465424,
0.0325985811650753,
-0.02209874987602234,
-0.22261080145835876,
-0.022111384198069572,
-0.054368991404771805,
-0.04119992256164551,
0.12181728333234787,
0.04460592940449715,
0.04549892991781235,
0.05165698006749153,
-0.034082889556884766,
-0.02702464535832405,
0.022554239258170128,
0.08637815713882446,
0.0512956939637661,
0.04222460463643074,
0.10316844284534454,
-0.020570114254951477,
-0.039728425443172455,
0.0026703919284045696,
0.01637844555079937,
0.25827744603157043,
0.0017435584450140595,
0.1831115484237671,
0.053666867315769196,
0.18405140936374664,
0.01713571883738041,
0.06738167256116867,
0.008248529396951199,
-0.027492046356201172,
0.03271493315696716,
-0.06292600929737091,
-0.04024316370487213,
0.061550937592983246,
0.047226954251527786,
0.06244306638836861,
-0.12787450850009918,
0.001458121812902391,
0.028056593611836433,
0.3435657322406769,
0.04557841643691063,
-0.2845531702041626,
-0.10086248070001602,
0.010196896269917488,
-0.08398351073265076,
-0.05867261067032814,
0.029195310547947884,
0.12144885957241058,
-0.10768278688192368,
0.052733223885297775,
-0.08975588530302048,
0.08538487553596497,
-0.028809022158384323,
0.00031226692954078317,
0.039060622453689575,
0.08592190593481064,
-0.0070873890072107315,
0.07601281255483627,
-0.27739477157592773,
0.2788859009742737,
-0.01931467279791832,
0.10148321092128754,
-0.05814915522933006,
0.01951705664396286,
0.059694595634937286,
-0.03958654776215553,
0.08038774877786636,
-0.013860291801393032,
-0.12730133533477783,
-0.17215926945209503,
-0.08034028857946396,
0.03201693668961525,
0.13056541979312897,
-0.05871972441673279,
0.12612399458885193,
-0.03841419517993927,
-0.0064153606072068214,
0.05346996337175369,
-0.0724182277917862,
-0.1252678483724594,
-0.08070661872625351,
0.02363334782421589,
0.016956908628344536,
0.07671157270669937,
-0.151870459318161,
-0.10835766047239304,
-0.06349948048591614,
0.08820343762636185,
-0.09676878154277802,
-0.023159347474575043,
-0.13435404002666473,
0.06577644497156143,
0.14648091793060303,
-0.07922086864709854,
0.053520530462265015,
0.027346450835466385,
0.1533922702074051,
0.011156944558024406,
-0.012929501011967659,
0.07914771884679794,
-0.08439458906650543,
-0.21155992150306702,
-0.0389849953353405,
0.18174247443675995,
0.04869459569454193,
0.05377953499555588,
0.0054792906157672405,
0.0077507710084319115,
-0.016493791714310646,
-0.08361169695854187,
0.039523884654045105,
0.008108694106340408,
-0.017501940950751305,
0.02396305464208126,
-0.0335964672267437,
-0.010915680788457394,
-0.0869302824139595,
-0.06060989573597908,
0.14321552217006683,
0.29109564423561096,
-0.06061743199825287,
0.015667326748371124,
0.055860862135887146,
-0.04084104299545288,
-0.16994450986385345,
0.007319422904402018,
0.1245623454451561,
0.04356773942708969,
-0.027241894975304604,
-0.20189210772514343,
0.053657956421375275,
0.06798949092626572,
-0.03478918969631195,
0.08207961171865463,
-0.28720006346702576,
-0.1480947583913803,
0.13514597713947296,
0.13468937575817108,
0.0005424652481451631,
-0.15908360481262207,
-0.07571447640657425,
-0.01240311749279499,
-0.05925225466489792,
0.07081178575754166,
-0.07403858751058578,
0.11065145581960678,
-0.004715829622000456,
0.043145377188920975,
0.029693644493818283,
-0.054808974266052246,
0.16105978190898895,
-0.019762935116887093,
0.06194566562771797,
-0.011050584726035595,
0.043991558253765106,
0.05051959678530693,
-0.0726996660232544,
0.024814819917082787,
-0.09497933834791183,
0.028163766488432884,
-0.12757080793380737,
-0.03166157007217407,
-0.09659477323293686,
0.04386800155043602,
-0.04240944981575012,
-0.03619907796382904,
-0.007944834418594837,
0.06203160434961319,
0.06585849821567535,
0.022258801385760307,
0.1250561773777008,
-0.053487155586481094,
0.1346907615661621,
0.0846765860915184,
0.12213544547557831,
-0.005320806987583637,
-0.048052262514829636,
-0.02153538353741169,
-0.011516709811985493,
0.05977892875671387,
-0.12186963856220245,
0.03967038542032242,
0.14190395176410675,
0.05028009042143822,
0.1489778310060501,
0.050591740757226944,
-0.11412771791219711,
0.013002557680010796,
0.07319384813308716,
-0.08512633293867111,
-0.14458471536636353,
-0.04802543297410011,
0.13115915656089783,
-0.13911831378936768,
0.02927277237176895,
0.09910706430673599,
-0.056366585195064545,
-0.011473785154521465,
-0.01577385514974594,
-0.0015592511044815183,
-0.04418018460273743,
0.2169143110513687,
0.0771903246641159,
0.09165140986442566,
-0.07675863057374954,
0.08037775754928589,
0.0301134642213583,
-0.0858350545167923,
0.035684213042259216,
0.08198612183332443,
-0.059903424233198166,
-0.023809446021914482,
-0.0013315830146893859,
0.09039702266454697,
0.013310679234564304,
-0.07761600613594055,
-0.13344071805477142,
-0.1363591104745865,
0.04047997295856476,
0.16191458702087402,
0.04244725778698921,
0.015142209827899933,
-0.017995044589042664,
0.03677058964967728,
-0.10728756338357925,
0.11228200793266296,
0.09771298617124557,
0.07949118316173553,
-0.14908020198345184,
0.14051011204719543,
0.008070722222328186,
0.024587051942944527,
-0.006240000948309898,
-0.014097368344664574,
-0.10295916348695755,
0.028112413361668587,
-0.14419691264629364,
0.0013625420397147536,
-0.05808153748512268,
-0.01143976952880621,
0.012623738497495651,
-0.07603080570697784,
-0.05360105633735657,
0.04039343073964119,
-0.11826664209365845,
-0.03709382563829422,
0.0017380116041749716,
0.0684996023774147,
-0.08725710958242416,
-0.040631234645843506,
0.05675508454442024,
-0.12629945576190948,
0.08985166251659393,
0.05740285664796829,
0.02431432157754898,
0.07755546271800995,
-0.1078559011220932,
0.02694249339401722,
0.0156581811606884,
0.008747223764657974,
0.006081452127546072,
-0.1571422815322876,
-0.01142306812107563,
-0.024828407913446426,
0.025327593088150024,
-0.011177832260727882,
0.023287959396839142,
-0.11984800547361374,
-0.043464791029691696,
-0.030566541478037834,
-0.02895781770348549,
-0.0635605901479721,
0.0528271347284317,
0.03262682631611824,
0.05563826858997345,
0.1513635218143463,
-0.09820006787776947,
0.028148848563432693,
-0.20930877327919006,
0.018517035990953445,
-0.03187176585197449,
-0.09429725259542465,
-0.038216911256313324,
-0.025193413719534874,
0.07465670257806778,
-0.07082703709602356,
0.05115460976958275,
-0.08713415265083313,
0.024620426818728447,
0.019014716148376465,
-0.0994972214102745,
0.055374741554260254,
0.05393773317337036,
0.2842618525028229,
0.03210046887397766,
-0.00045735063031315804,
0.07342847436666489,
-0.0037523964419960976,
0.043577760457992554,
0.12146712839603424,
0.13489723205566406,
0.19119107723236084,
0.030238915234804153,
0.0946178212761879,
0.08005556464195251,
-0.0808810293674469,
-0.11139720678329468,
0.10621180385351181,
-0.02511557564139366,
0.0791650339961052,
-0.010110096074640751,
0.21773362159729004,
0.1555735319852829,
-0.19165153801441193,
0.029818153008818626,
-0.0221396517008543,
-0.08811792731285095,
-0.08693249523639679,
-0.09950625151395798,
-0.08046285063028336,
-0.1455574482679367,
0.017505424097180367,
-0.10574012249708176,
0.023806635290384293,
0.05735178664326668,
0.03682541847229004,
0.015793753787875175,
0.16645461320877075,
0.04041777923703194,
-0.006275073625147343,
0.11894796043634415,
-0.01215969305485487,
-0.03343753144145012,
-0.019683517515659332,
-0.11581163853406906,
0.04799019172787666,
-0.03413540869951248,
0.040217939764261246,
-0.057159826159477234,
-0.0819007009267807,
0.054813798516988754,
0.01074721198529005,
-0.10027574002742767,
0.022143887355923653,
-0.014738288708031178,
0.07531974464654922,
0.0607801154255867,
0.046650879085063934,
-0.02024996094405651,
-0.019661802798509598,
0.23358917236328125,
-0.0795273706316948,
-0.08146709948778152,
-0.1343500167131424,
0.20922785997390747,
0.017693040892481804,
-0.027523653581738472,
0.025653459131717682,
-0.07914656400680542,
0.03977824002504349,
0.15344025194644928,
0.09217271953821182,
-0.015276536345481873,
0.004110413137823343,
-0.015447444282472134,
-0.016720598563551903,
-0.07598797231912613,
0.08111870288848877,
0.10238002240657806,
0.02286483719944954,
-0.06460048258304596,
-0.035003017634153366,
-0.03426097333431244,
-0.04404864087700844,
-0.054579783231019974,
0.0907328724861145,
0.02468237280845642,
-0.0168245080858469,
-0.03748966380953789,
0.09766615927219391,
-0.06135124713182449,
-0.08435352891683578,
0.0012740462552756071,
-0.14297810196876526,
-0.17475292086601257,
-0.04674718901515007,
0.024847829714417458,
0.025061843916773796,
0.03185536712408066,
-0.019966959953308105,
-0.0014520654221996665,
0.055910930037498474,
0.012510457076132298,
-0.020705407485365868,
-0.08160096406936646,
0.11027306318283081,
-0.09710359573364258,
0.22480374574661255,
-0.05440472811460495,
0.023724887520074844,
0.13460499048233032,
0.04197513312101364,
-0.08278360217809677,
0.06893691420555115,
0.07356136292219162,
-0.12855876982212067,
0.04141200706362724,
0.20273752510547638,
-0.030594514682888985,
0.15903215110301971,
0.026858512312173843,
-0.11229404807090759,
0.01120186410844326,
-0.06082075834274292,
-0.04380065202713013,
-0.06857775151729584,
0.022008737549185753,
-0.05601683259010315,
0.13962528109550476,
0.17315030097961426,
-0.0914076417684555,
-0.01782986894249916,
-0.060665108263492584,
0.02934652753174305,
0.05402298644185066,
0.10261973738670349,
-0.03333732485771179,
-0.2813875079154968,
0.01311715878546238,
-0.021609380841255188,
-0.009802333079278469,
-0.24957215785980225,
-0.09234141558408737,
0.011853920295834541,
-0.05802639573812485,
-0.04265066236257553,
0.11975325644016266,
0.11260280013084412,
0.046010375022888184,
-0.04737485200166702,
-0.08593881875276566,
-0.06596330553293228,
0.17762155830860138,
-0.17229454219341278,
-0.06784893572330475
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SMIDS_3x_beit_large_RMSProp_lr0001_fold1
This model is a fine-tuned version of [microsoft/beit-large-patch16-224](https://huggingface.co/microsoft/beit-large-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.1638
- Accuracy: 0.8848
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.467 | 1.0 | 451 | 0.4114 | 0.8230 |
| 0.3239 | 2.0 | 902 | 0.3424 | 0.8648 |
| 0.2765 | 3.0 | 1353 | 0.4139 | 0.8531 |
| 0.2282 | 4.0 | 1804 | 0.3451 | 0.8865 |
| 0.1605 | 5.0 | 2255 | 0.4701 | 0.8748 |
| 0.1651 | 6.0 | 2706 | 0.4789 | 0.8798 |
| 0.1547 | 7.0 | 3157 | 0.4952 | 0.8815 |
| 0.1301 | 8.0 | 3608 | 0.4426 | 0.8715 |
| 0.1681 | 9.0 | 4059 | 0.5755 | 0.8681 |
| 0.0267 | 10.0 | 4510 | 0.6593 | 0.8698 |
| 0.0932 | 11.0 | 4961 | 0.4949 | 0.8715 |
| 0.0591 | 12.0 | 5412 | 0.6634 | 0.8748 |
| 0.0288 | 13.0 | 5863 | 0.5789 | 0.9015 |
| 0.0018 | 14.0 | 6314 | 0.6155 | 0.8881 |
| 0.0273 | 15.0 | 6765 | 0.7024 | 0.8698 |
| 0.0311 | 16.0 | 7216 | 0.7733 | 0.8581 |
| 0.0203 | 17.0 | 7667 | 0.7893 | 0.8765 |
| 0.0002 | 18.0 | 8118 | 0.9239 | 0.8798 |
| 0.0465 | 19.0 | 8569 | 0.6952 | 0.8881 |
| 0.0769 | 20.0 | 9020 | 0.7171 | 0.8865 |
| 0.0014 | 21.0 | 9471 | 0.8100 | 0.8715 |
| 0.0013 | 22.0 | 9922 | 0.7003 | 0.8765 |
| 0.0277 | 23.0 | 10373 | 0.8631 | 0.8781 |
| 0.0002 | 24.0 | 10824 | 0.9872 | 0.8765 |
| 0.0001 | 25.0 | 11275 | 0.7627 | 0.8948 |
| 0.0331 | 26.0 | 11726 | 0.8254 | 0.8915 |
| 0.0061 | 27.0 | 12177 | 0.8133 | 0.8932 |
| 0.0187 | 28.0 | 12628 | 0.9134 | 0.8765 |
| 0.0002 | 29.0 | 13079 | 0.9734 | 0.8831 |
| 0.0172 | 30.0 | 13530 | 0.7547 | 0.8831 |
| 0.0 | 31.0 | 13981 | 0.8396 | 0.8865 |
| 0.0001 | 32.0 | 14432 | 1.0245 | 0.8815 |
| 0.0 | 33.0 | 14883 | 0.7812 | 0.9015 |
| 0.0055 | 34.0 | 15334 | 0.9777 | 0.8965 |
| 0.0062 | 35.0 | 15785 | 1.1854 | 0.8831 |
| 0.0 | 36.0 | 16236 | 0.9153 | 0.8881 |
| 0.0 | 37.0 | 16687 | 1.0422 | 0.8898 |
| 0.0 | 38.0 | 17138 | 1.1991 | 0.8798 |
| 0.0073 | 39.0 | 17589 | 0.9200 | 0.8982 |
| 0.0 | 40.0 | 18040 | 0.9841 | 0.9015 |
| 0.0 | 41.0 | 18491 | 1.0684 | 0.8982 |
| 0.0104 | 42.0 | 18942 | 1.0722 | 0.9032 |
| 0.0 | 43.0 | 19393 | 1.0249 | 0.8765 |
| 0.0 | 44.0 | 19844 | 1.0517 | 0.8848 |
| 0.0 | 45.0 | 20295 | 1.1233 | 0.8831 |
| 0.0 | 46.0 | 20746 | 1.1565 | 0.8848 |
| 0.0 | 47.0 | 21197 | 1.1511 | 0.8815 |
| 0.0 | 48.0 | 21648 | 1.1693 | 0.8798 |
| 0.0 | 49.0 | 22099 | 1.1569 | 0.8865 |
| 0.0 | 50.0 | 22550 | 1.1638 | 0.8848 |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1
- Datasets 2.12.0
- Tokenizers 0.13.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/beit-large-patch16-224", "model-index": [{"name": "SMIDS_3x_beit_large_RMSProp_lr0001_fold1", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "test", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.8848080133555927, "name": "Accuracy"}]}]}]} | image-classification | onizukal/SMIDS_3x_beit_large_RMSProp_lr0001_fold1 | [
"transformers",
"pytorch",
"beit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/beit-large-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:13:25+00:00 | [] | [] | TAGS
#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| SMIDS\_3x\_beit\_large\_RMSProp\_lr0001\_fold1
==============================================
This model is a fine-tuned version of microsoft/beit-large-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.1638
* Accuracy: 0.8848
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0001
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.32.1
* Pytorch 2.0.1
* Datasets 2.12.0
* Tokenizers 0.13.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
"TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
81,
115,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
-0.12968555092811584,
0.17251011729240417,
-0.0023243443574756384,
0.1362919956445694,
0.1120586097240448,
0.015268749557435513,
0.14003369212150574,
0.16890837252140045,
-0.08239254355430603,
0.046998485922813416,
0.14023225009441376,
0.13628867268562317,
0.046756189316511154,
0.19432850182056427,
-0.052493587136268616,
-0.26022207736968994,
0.04113864526152611,
0.032812196761369705,
-0.020441479980945587,
0.1235608458518982,
0.09337224811315536,
-0.13087525963783264,
0.11667836457490921,
0.0301132183521986,
-0.20004093647003174,
-0.036873914301395416,
-0.007245634216815233,
-0.06722474098205566,
0.10533155500888824,
-0.0034045001957565546,
0.0691065788269043,
0.03768180310726166,
0.08387713134288788,
-0.13018712401390076,
0.002076903358101845,
0.042768821120262146,
0.0062860166653990746,
0.10383369028568268,
0.054196570068597794,
-0.015545758418738842,
0.0701410248875618,
-0.06851525604724884,
0.0672622099518776,
0.009240911342203617,
-0.11321496963500977,
-0.2700493633747101,
-0.10203396528959274,
0.07240316271781921,
0.08221714198589325,
0.06822962313890457,
0.008172801695764065,
0.16417047381401062,
-0.014714903198182583,
0.10454332083463669,
0.23100516200065613,
-0.26415953040122986,
-0.05532161891460419,
0.029576225206255913,
0.015004046261310577,
0.06490366160869598,
-0.10617698729038239,
-0.01859438419342041,
0.020827138796448708,
0.04436356946825981,
0.1411312073469162,
-0.010821618139743805,
-0.028378209099173546,
-0.021572042256593704,
-0.10856294631958008,
-0.08875563740730286,
0.18566860258579254,
0.05809066444635391,
-0.048288628458976746,
-0.07735078781843185,
-0.07127056270837784,
-0.17220835387706757,
-0.041861895471811295,
0.009548050351440907,
0.041730549186468124,
-0.04684269055724144,
-0.10686429589986801,
-0.031055882573127747,
-0.078252874314785,
-0.051669858396053314,
-0.023303553462028503,
0.13525931537151337,
0.03357808664441109,
0.05729198828339577,
-0.03593141585588455,
0.09915280342102051,
0.006841922644525766,
-0.17527513206005096,
-0.028045548126101494,
-0.0016165260458365083,
0.01563161052763462,
-0.020048104226589203,
-0.03057136945426464,
-0.06562764942646027,
-0.0016239769756793976,
0.149040088057518,
-0.06106079742312431,
0.06079873815178871,
-0.0069216229021549225,
0.04031313583254814,
-0.0486484132707119,
0.18668954074382782,
-0.028643600642681122,
-0.016713637858629227,
0.02057800441980362,
0.08857519924640656,
0.06818821281194687,
-0.03644402697682381,
-0.12566283345222473,
0.03087625838816166,
0.1283741444349289,
0.0027549222577363253,
-0.021953243762254715,
0.053039632737636566,
-0.06444176286458969,
-0.05842158570885658,
0.09141092747449875,
-0.08884678035974503,
0.03514961525797844,
-0.01055920124053955,
-0.08416686952114105,
-0.06807748228311539,
0.02709859050810337,
0.018840007483959198,
-0.00014874596672598273,
0.07201956957578659,
-0.09116632491350174,
0.015490563586354256,
-0.06551176309585571,
-0.10091431438922882,
0.01564670167863369,
-0.11040772497653961,
0.012323775328695774,
-0.09688954800367355,
-0.1969451904296875,
0.006960712838917971,
0.07738039642572403,
-0.05607226490974426,
-0.06792453676462173,
-0.03661259636282921,
-0.07637017965316772,
0.04143770784139633,
-0.01186586357653141,
0.07317496836185455,
-0.07456725090742111,
0.09119440615177155,
0.02237127535045147,
0.08760105073451996,
-0.056383248418569565,
0.04597126320004463,
-0.10241573303937912,
0.04992371052503586,
-0.19877833127975464,
0.07988634705543518,
-0.049189720302820206,
0.06190093979239464,
-0.09581396728754044,
-0.10568851977586746,
0.033553607761859894,
-0.04994693025946617,
0.068512924015522,
0.09739063680171967,
-0.17317676544189453,
-0.05787286534905434,
0.13517500460147858,
-0.09691634029150009,
-0.14840039610862732,
0.10115666687488556,
-0.05093328654766083,
0.019768450409173965,
0.04739697277545929,
0.21447287499904633,
0.062935970723629,
-0.0910891741514206,
-0.025994082912802696,
-0.03333966061472893,
0.044677652418613434,
-0.06483115255832672,
0.101903036236763,
0.027484174817800522,
0.0531504862010479,
0.02367355115711689,
-0.03332329913973808,
0.03818739578127861,
-0.08385370671749115,
-0.10085898637771606,
-0.05038752406835556,
-0.08557170629501343,
0.039683446288108826,
0.05594057962298393,
0.059847064316272736,
-0.10873348265886307,
-0.09023979306221008,
0.041734639555215836,
0.09406744688749313,
-0.07396076619625092,
0.02903648279607296,
-0.0904788002371788,
0.11622294038534164,
-0.08363831788301468,
-0.02404896728694439,
-0.17903628945350647,
-0.0417308546602726,
0.04055763781070709,
-0.01668366603553295,
-0.006775525398552418,
-0.0494389571249485,
0.07092705368995667,
0.087753064930439,
-0.05281677842140198,
-0.052284084260463715,
-0.05530114471912384,
0.008562305010855198,
-0.11059658974409103,
-0.1778055727481842,
-0.080107681453228,
-0.03797448053956032,
0.15019145607948303,
-0.15246915817260742,
0.0224970243871212,
0.0616903156042099,
0.12470164895057678,
0.05992257222533226,
-0.0469760037958622,
-0.007631834130734205,
0.0217386856675148,
-0.05561714619398117,
-0.0865136981010437,
0.05727535858750343,
0.035165008157491684,
-0.07172347605228424,
-0.019373787567019463,
-0.10040221363306046,
0.15015454590320587,
0.13185308873653412,
-0.0021352346520870924,
-0.045590728521347046,
-0.012053865939378738,
-0.06572475284337997,
-0.030354894697666168,
-0.04096601903438568,
0.01860888861119747,
0.1020345464348793,
0.017360014840960503,
0.14407898485660553,
-0.09213681519031525,
-0.037007302045822144,
0.053231216967105865,
-0.028658904135227203,
-0.03313332051038742,
0.0737093985080719,
0.021478038281202316,
-0.14289474487304688,
0.1502111405134201,
0.14915579557418823,
-0.04949729144573212,
0.12371271848678589,
-0.03663388267159462,
-0.06141006201505661,
-0.04545919969677925,
-0.03777514770627022,
0.01429951936006546,
0.1407921016216278,
-0.08363746106624603,
-0.006257671397179365,
0.05626929551362991,
0.018998416140675545,
-0.007220869418233633,
-0.1808812916278839,
0.0005758196348324418,
0.03530525416135788,
-0.04614398628473282,
-0.022574707865715027,
-0.014720434322953224,
0.000520858506206423,
0.09188775718212128,
0.02001834660768509,
-0.07113038748502731,
0.05185159295797348,
0.010694033466279507,
-0.056145116686820984,
0.16459684073925018,
-0.07884351164102554,
-0.19753409922122955,
-0.11793240904808044,
-0.08745986223220825,
-0.10736268758773804,
0.013000035658478737,
0.067270427942276,
-0.050670597702264786,
-0.04932181537151337,
-0.1026671901345253,
-0.044550344347953796,
0.021845674142241478,
0.024347107857465744,
0.053595975041389465,
-0.00796813890337944,
0.08411940932273865,
-0.09194666892290115,
-0.03317512199282646,
-0.014813165180385113,
0.01894056238234043,
0.0670066773891449,
0.01914203353226185,
0.11091019958257675,
0.08160436898469925,
-0.0286879725754261,
0.05666669085621834,
-0.01685662567615509,
0.26526889204978943,
-0.06748054921627045,
-0.006749235559254885,
0.1391732543706894,
-0.013490693643689156,
0.0842166393995285,
0.12729591131210327,
0.04176322743296623,
-0.0955888107419014,
-0.01310211792588234,
-0.0005005627172067761,
-0.05257550999522209,
-0.1536482274532318,
-0.04132819548249245,
-0.04548354819417,
-0.0018228141125291586,
0.13951772451400757,
0.038064174354076385,
0.02505229413509369,
0.07843583822250366,
0.020602436736226082,
0.05678323283791542,
-0.0175874512642622,
0.10429482907056808,
0.08156884461641312,
0.06449971348047256,
0.13376133143901825,
-0.036523740738630295,
-0.019790813326835632,
0.05638623237609863,
0.042081572115421295,
0.20467498898506165,
-0.025362396612763405,
0.14717818796634674,
0.026553483679890633,
0.19327539205551147,
0.017808275297284126,
0.07306244969367981,
-0.014873637817800045,
0.0007499073399230838,
-0.019323905929923058,
-0.04713669419288635,
-0.0638502836227417,
0.03312433883547783,
-0.016851995140314102,
0.05682634562253952,
-0.09328699111938477,
0.03906902298331261,
0.05959288775920868,
0.30634987354278564,
0.0654144361615181,
-0.4125381410121918,
-0.09821337461471558,
0.012344546616077423,
0.0008716733427718282,
-0.05509618669748306,
-0.007402430288493633,
0.0980701595544815,
-0.09973937273025513,
0.0819711834192276,
-0.09416680037975311,
0.08507230132818222,
-0.0846736952662468,
0.020382488146424294,
0.07683569937944412,
0.055889930576086044,
0.012921135872602463,
0.05964238941669464,
-0.21880683302879333,
0.2499670386314392,
0.01837102696299553,
0.04415145888924599,
-0.08875706046819687,
0.009965145029127598,
0.03320525959134102,
0.05923061817884445,
0.08590700477361679,
0.0061045982874929905,
-0.09025654941797256,
-0.18889141082763672,
-0.12562422454357147,
0.000394518458051607,
0.06176565960049629,
-0.03729195147752762,
0.09444484859704971,
-0.018019067123532295,
-0.012201022356748581,
0.02127370797097683,
0.0009904175531119108,
-0.035084888339042664,
-0.10356581956148148,
0.02010609768331051,
0.03430531173944473,
-0.011726552620530128,
-0.06489048153162003,
-0.11480618268251419,
-0.035277001559734344,
0.16168422996997833,
0.05518770217895508,
-0.07543513178825378,
-0.14076673984527588,
0.0721859410405159,
0.0775376707315445,
-0.08563373237848282,
0.03936640918254852,
-0.016648126766085625,
0.14995604753494263,
0.020845195278525352,
-0.0889848992228508,
0.10199198871850967,
-0.05838112160563469,
-0.17863209545612335,
-0.04141612723469734,
0.09901762008666992,
0.007052883040159941,
0.05273612216114998,
0.004226623103022575,
0.06022334843873978,
-0.03518751636147499,
-0.05844981223344803,
0.06672939658164978,
-0.007545650005340576,
0.10645230114459991,
-0.014578265137970448,
0.008669902570545673,
0.028680432587862015,
-0.046410609036684036,
0.00012374592188280076,
0.1686571091413498,
0.24114695191383362,
-0.10427109152078629,
0.060499124228954315,
0.03038850799202919,
-0.030858036130666733,
-0.18259160220623016,
0.01086394116282463,
0.07622820883989334,
-0.00013084696547593921,
0.04143662750720978,
-0.1601918637752533,
0.05532059073448181,
0.10498367995023727,
-0.043228019028902054,
0.08107142895460129,
-0.27694207429885864,
-0.1185181736946106,
0.09238865971565247,
0.13856256008148193,
0.06877914071083069,
-0.13106170296669006,
-0.043299052864313126,
-0.041688259690999985,
-0.17338812351226807,
0.13653364777565002,
-0.057192787528038025,
0.1145344004034996,
-0.039500072598457336,
0.08082033693790436,
0.014952262863516808,
-0.056017596274614334,
0.14574900269508362,
0.0056154001504182816,
0.08686088770627975,
-0.07213473320007324,
-0.0020430299919098616,
0.10663212835788727,
-0.10254329442977905,
0.07232339680194855,
-0.08735590428113937,
0.0618043914437294,
-0.10790637135505676,
-0.003900582902133465,
-0.07402003556489944,
0.013697824440896511,
-0.01366274245083332,
-0.04917207732796669,
-0.04516566917300224,
0.03515308350324631,
0.0627121776342392,
-0.01822420209646225,
0.20940853655338287,
0.06430324167013168,
0.08635561168193817,
0.1727360188961029,
0.054769597947597504,
-0.10558480769395828,
-0.09403572231531143,
-0.043973103165626526,
-0.029537810012698174,
0.05986782908439636,
-0.1372820883989334,
0.0528247207403183,
0.11996810883283615,
0.013451187871396542,
0.12858225405216217,
0.055897701531648636,
-0.030677761882543564,
0.03560479357838631,
0.062153734266757965,
-0.17216050624847412,
-0.08662130683660507,
-0.009840693324804306,
0.030872231349349022,
-0.13055209815502167,
0.0458756685256958,
0.12116101384162903,
-0.05953402817249298,
-0.015017039142549038,
-0.004467411432415247,
0.03673877567052841,
-0.00978675577789545,
0.15920081734657288,
0.048089753836393356,
0.055168475955724716,
-0.11802823096513748,
0.11332250386476517,
0.05730176344513893,
-0.07302459329366684,
0.03206014260649681,
0.05020790174603462,
-0.1039617657661438,
-0.021727759391069412,
0.03114185482263565,
0.15037071704864502,
-0.06283780187368393,
-0.045329563319683075,
-0.1358855813741684,
-0.09226331859827042,
0.06643375009298325,
0.07981554418802261,
0.09349396824836731,
0.016502337530255318,
-0.03525979816913605,
-0.013309485279023647,
-0.10845191776752472,
0.11000601947307587,
0.04338005557656288,
0.09121100604534149,
-0.17974577844142914,
0.05434896796941757,
-0.001805671607144177,
0.07240304350852966,
-0.02173563651740551,
-0.00018242778605781496,
-0.08797106891870499,
0.0035262287128716707,
-0.10818753391504288,
0.024682866409420967,
-0.052850391715765,
0.006376184988766909,
-0.020511267706751823,
-0.05819518491625786,
-0.06372886151075363,
0.024663057178258896,
-0.1193968653678894,
-0.05304655060172081,
0.02193489298224449,
0.03176874667406082,
-0.11983832716941833,
-0.04395153746008873,
0.02043171599507332,
-0.08966860175132751,
0.09786758571863174,
0.06017395853996277,
-0.00797541905194521,
0.007467431016266346,
0.0038150406908243895,
-0.022212069481611252,
0.06630469858646393,
0.0074848150834441185,
0.08584009110927582,
-0.11553936451673508,
-0.022143544629216194,
0.016299601644277573,
-0.004447818733751774,
0.018147116526961327,
0.1585858017206192,
-0.12092386186122894,
0.00018621055642142892,
-0.014765054918825626,
-0.06592588871717453,
-0.06358986347913742,
0.0692417323589325,
0.10919524729251862,
0.02367839775979519,
0.2122299075126648,
-0.054594267159700394,
0.015877852216362953,
-0.21000300347805023,
-0.011462570168077946,
0.005311926826834679,
-0.13887609541416168,
-0.10537440329790115,
-0.032787878066301346,
0.0637630894780159,
-0.07039659470319748,
0.1177176982164383,
0.03537357598543167,
0.020886771380901337,
0.02911887876689434,
0.024869181215763092,
-0.002677198965102434,
0.013766518794000149,
0.1633930504322052,
0.014011929742991924,
-0.02872646041214466,
0.1283825933933258,
0.029096294194459915,
0.09337089955806732,
0.11805824935436249,
0.1763046532869339,
0.11451227962970734,
0.0477789007127285,
0.09043081104755402,
0.0520024336874485,
-0.02513159066438675,
-0.22147811949253082,
0.036259569227695465,
-0.039764102548360825,
0.1483127623796463,
-0.0033327124547213316,
0.15980194509029388,
0.09223487228155136,
-0.18392090499401093,
0.040660299360752106,
-0.037005215883255005,
-0.07937940210103989,
-0.08421849459409714,
-0.12178675830364227,
-0.1033017709851265,
-0.1509413868188858,
0.0028559700585901737,
-0.10428426414728165,
0.022927863523364067,
0.11217869818210602,
-0.008710348978638649,
-0.010019375011324883,
0.11695955693721771,
-0.026584560051560402,
0.026202335953712463,
0.03870072960853577,
0.00616151699796319,
-0.05987776443362236,
-0.04411191865801811,
-0.08036603778600693,
0.014018801040947437,
0.03200533241033554,
0.055842287838459015,
-0.03226681798696518,
-0.007200593128800392,
0.03782269358634949,
-0.009845683351159096,
-0.12363012880086899,
0.013544945046305656,
0.004753641318529844,
0.05189259722828865,
0.0008605605689808726,
0.01290043629705906,
0.03187544271349907,
-0.015199882909655571,
0.193119078874588,
-0.07321906089782715,
-0.02744952403008938,
-0.12274995446205139,
0.17869888246059418,
0.0023205638863146305,
-0.049724213778972626,
0.05292708799242973,
-0.09127075970172882,
-0.020290102809667587,
0.1547212302684784,
0.18941837549209595,
-0.07176556438207626,
-0.01638839766383171,
-0.017501909285783768,
-0.01388427522033453,
-0.022741587832570076,
0.09889717400074005,
0.09887372702360153,
-0.007504772394895554,
-0.07518953084945679,
-0.028498217463493347,
-0.06611054390668869,
-0.03444022685289383,
-0.03838160261511803,
0.06909165531396866,
-0.004605968948453665,
0.007089514285326004,
-0.0751754567027092,
0.04334408789873123,
-0.02207781746983528,
-0.060899440199136734,
0.06262887269258499,
-0.21282166242599487,
-0.17796695232391357,
0.006926008500158787,
0.07579630613327026,
0.0016649233875796199,
0.04621230810880661,
-0.010005760937929153,
0.018681904301047325,
0.07549776136875153,
-0.022177988663315773,
-0.0866948589682579,
-0.09604813903570175,
0.1083223819732666,
-0.1344224065542221,
0.25299492478370667,
-0.03893125429749489,
0.035907670855522156,
0.12175600975751877,
0.041717030107975006,
-0.13353091478347778,
0.033571965992450714,
0.03969275578856468,
-0.03212675452232361,
0.005746500100940466,
0.14248594641685486,
-0.037242501974105835,
0.07988674938678741,
0.04599026218056679,
-0.10243327170610428,
-0.039464809000492096,
-0.04960913211107254,
-0.011240639723837376,
-0.024744588881731033,
-0.05439573898911476,
-0.03649099916219711,
0.13208730518817902,
0.17168967425823212,
-0.04232889041304588,
-0.023784559220075607,
-0.06460724771022797,
0.030773790553212166,
0.0774260088801384,
-0.033050306141376495,
-0.05197038874030113,
-0.23585109412670135,
0.0024181774351745844,
0.05249672383069992,
-0.013345940038561821,
-0.20789918303489685,
-0.11062979698181152,
0.006115853786468506,
-0.05795856565237045,
-0.07630864530801773,
0.09230074286460876,
0.06326484680175781,
0.035358402878046036,
-0.06319575011730194,
0.03810267895460129,
-0.07874377071857452,
0.1419457346200943,
-0.1448507308959961,
-0.07860494405031204
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SMIDS_3x_beit_large_RMSProp_lr001_fold1
This model is a fine-tuned version of [microsoft/beit-large-patch16-224](https://huggingface.co/microsoft/beit-large-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.9451
- Accuracy: 0.7846
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.9818 | 1.0 | 451 | 1.0065 | 0.5042 |
| 0.8866 | 2.0 | 902 | 0.8036 | 0.5543 |
| 0.8057 | 3.0 | 1353 | 0.7826 | 0.6210 |
| 0.782 | 4.0 | 1804 | 0.8406 | 0.5893 |
| 0.7177 | 5.0 | 2255 | 0.7766 | 0.6294 |
| 0.714 | 6.0 | 2706 | 0.8601 | 0.6110 |
| 0.7301 | 7.0 | 3157 | 0.7598 | 0.6244 |
| 0.7096 | 8.0 | 3608 | 0.7396 | 0.6411 |
| 0.674 | 9.0 | 4059 | 0.7122 | 0.6511 |
| 0.6528 | 10.0 | 4510 | 0.7225 | 0.6578 |
| 0.7275 | 11.0 | 4961 | 0.7112 | 0.6644 |
| 0.7084 | 12.0 | 5412 | 0.7121 | 0.6678 |
| 0.6609 | 13.0 | 5863 | 0.7224 | 0.6611 |
| 0.6102 | 14.0 | 6314 | 0.6504 | 0.7162 |
| 0.6754 | 15.0 | 6765 | 0.6465 | 0.7346 |
| 0.6619 | 16.0 | 7216 | 0.7280 | 0.6828 |
| 0.5585 | 17.0 | 7667 | 0.6490 | 0.7362 |
| 0.5824 | 18.0 | 8118 | 0.6272 | 0.7329 |
| 0.5641 | 19.0 | 8569 | 0.7113 | 0.6795 |
| 0.5807 | 20.0 | 9020 | 0.6510 | 0.7312 |
| 0.4625 | 21.0 | 9471 | 0.6480 | 0.7412 |
| 0.4992 | 22.0 | 9922 | 0.6230 | 0.7412 |
| 0.4944 | 23.0 | 10373 | 0.6202 | 0.7730 |
| 0.5334 | 24.0 | 10824 | 0.6215 | 0.7629 |
| 0.495 | 25.0 | 11275 | 0.6512 | 0.7429 |
| 0.4575 | 26.0 | 11726 | 0.6122 | 0.7629 |
| 0.4237 | 27.0 | 12177 | 0.5877 | 0.7713 |
| 0.501 | 28.0 | 12628 | 0.7039 | 0.7112 |
| 0.3935 | 29.0 | 13079 | 0.5859 | 0.7830 |
| 0.4914 | 30.0 | 13530 | 0.5515 | 0.7880 |
| 0.4409 | 31.0 | 13981 | 0.5910 | 0.7746 |
| 0.4164 | 32.0 | 14432 | 0.5925 | 0.7913 |
| 0.3075 | 33.0 | 14883 | 0.6194 | 0.7863 |
| 0.4528 | 34.0 | 15334 | 0.5905 | 0.7796 |
| 0.41 | 35.0 | 15785 | 0.5836 | 0.8047 |
| 0.3679 | 36.0 | 16236 | 0.5875 | 0.7963 |
| 0.3906 | 37.0 | 16687 | 0.6031 | 0.7963 |
| 0.3932 | 38.0 | 17138 | 0.6262 | 0.7830 |
| 0.3054 | 39.0 | 17589 | 0.6577 | 0.7846 |
| 0.2711 | 40.0 | 18040 | 0.6789 | 0.7980 |
| 0.3077 | 41.0 | 18491 | 0.6804 | 0.7830 |
| 0.2701 | 42.0 | 18942 | 0.7232 | 0.7980 |
| 0.3307 | 43.0 | 19393 | 0.7018 | 0.7796 |
| 0.2751 | 44.0 | 19844 | 0.7219 | 0.7846 |
| 0.2205 | 45.0 | 20295 | 0.7445 | 0.7880 |
| 0.2015 | 46.0 | 20746 | 0.8209 | 0.7880 |
| 0.2845 | 47.0 | 21197 | 0.8487 | 0.7796 |
| 0.3266 | 48.0 | 21648 | 0.9033 | 0.7746 |
| 0.2425 | 49.0 | 22099 | 0.9204 | 0.7863 |
| 0.1226 | 50.0 | 22550 | 0.9451 | 0.7846 |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1
- Datasets 2.12.0
- Tokenizers 0.13.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/beit-large-patch16-224", "model-index": [{"name": "SMIDS_3x_beit_large_RMSProp_lr001_fold1", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "test", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.7846410684474123, "name": "Accuracy"}]}]}]} | image-classification | onizukal/SMIDS_3x_beit_large_RMSProp_lr001_fold1 | [
"transformers",
"pytorch",
"beit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/beit-large-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:14:36+00:00 | [] | [] | TAGS
#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| SMIDS\_3x\_beit\_large\_RMSProp\_lr001\_fold1
=============================================
This model is a fine-tuned version of microsoft/beit-large-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.9451
* Accuracy: 0.7846
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.001
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.32.1
* Pytorch 2.0.1
* Datasets 2.12.0
* Tokenizers 0.13.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
"TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
81,
115,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
-0.1293599158525467,
0.1724882274866104,
-0.0023454553447663784,
0.13587836921215057,
0.11250235140323639,
0.015344180166721344,
0.13944171369075775,
0.16905122995376587,
-0.08231265097856522,
0.04725457355380058,
0.1399839073419571,
0.13659004867076874,
0.046719420701265335,
0.19427457451820374,
-0.05236957222223282,
-0.2601667046546936,
0.04119328409433365,
0.03234807401895523,
-0.02075078710913658,
0.12314869463443756,
0.09301083534955978,
-0.13055187463760376,
0.11641614139080048,
0.030263151973485947,
-0.1994711011648178,
-0.03690062463283539,
-0.00742433313280344,
-0.06729142367839813,
0.10523199290037155,
-0.003170925425365567,
0.06897550076246262,
0.03744976967573166,
0.0835329070687294,
-0.13024620711803436,
0.0019813377875834703,
0.043246712535619736,
0.0064185261726379395,
0.10353662818670273,
0.05471734330058098,
-0.015182994306087494,
0.07030405849218369,
-0.06875743716955185,
0.06702885776758194,
0.009385199286043644,
-0.11313743144273758,
-0.26980340480804443,
-0.10223843902349472,
0.07157823443412781,
0.08217991888523102,
0.068179190158844,
0.008332090452313423,
0.1646922081708908,
-0.015027978457510471,
0.10447341948747635,
0.23090173304080963,
-0.2640359699726105,
-0.055165741592645645,
0.0294360164552927,
0.014914325438439846,
0.06473758071660995,
-0.10603974759578705,
-0.018760167062282562,
0.02059783600270748,
0.044390130788087845,
0.1409236192703247,
-0.010635782033205032,
-0.02811739780008793,
-0.021928580477833748,
-0.10847067832946777,
-0.08875154703855515,
0.18579065799713135,
0.058072153478860855,
-0.04802494868636131,
-0.07736620306968689,
-0.07186304777860641,
-0.17165552079677582,
-0.041741833090782166,
0.009797174483537674,
0.04154014587402344,
-0.04674985632300377,
-0.10634559392929077,
-0.030910290777683258,
-0.07805538177490234,
-0.051461391150951385,
-0.023120464757084846,
0.1351369321346283,
0.03383360058069229,
0.05723920464515686,
-0.03597215935587883,
0.09929350018501282,
0.0073938071727752686,
-0.17543649673461914,
-0.028137801215052605,
-0.0017277015140280128,
0.015125435777008533,
-0.01991228759288788,
-0.030262885615229607,
-0.06523662805557251,
-0.001314454828388989,
0.1489848643541336,
-0.06082969531416893,
0.060913555324077606,
-0.007318461779505014,
0.04020942002534866,
-0.048562191426754,
0.1864238679409027,
-0.02870783396065235,
-0.01659870520234108,
0.0207351166754961,
0.08822518587112427,
0.06836054474115372,
-0.036532942205667496,
-0.12525734305381775,
0.03098621405661106,
0.12811045348644257,
0.0029374780133366585,
-0.021369412541389465,
0.05286439135670662,
-0.0643334686756134,
-0.0587083138525486,
0.09228596836328506,
-0.08908151835203171,
0.034838590770959854,
-0.010374085046350956,
-0.084370456635952,
-0.06779533624649048,
0.027354132384061813,
0.01850878819823265,
-0.0004323708708398044,
0.07165450602769852,
-0.09108522534370422,
0.014963540248572826,
-0.06533985584974289,
-0.10081785917282104,
0.016071073710918427,
-0.1107335016131401,
0.012424856424331665,
-0.09663169085979462,
-0.19710472226142883,
0.006891206838190556,
0.07721206545829773,
-0.05610070005059242,
-0.06814593821763992,
-0.03639180585741997,
-0.07652068138122559,
0.04154945909976959,
-0.011949662119150162,
0.07310567051172256,
-0.0747697651386261,
0.0913747176527977,
0.022405456751585007,
0.08734618872404099,
-0.05631003528833389,
0.0460691936314106,
-0.1024455726146698,
0.04984736442565918,
-0.19827407598495483,
0.0799676924943924,
-0.0493633896112442,
0.0617455318570137,
-0.09557844698429108,
-0.10542625933885574,
0.03370443359017372,
-0.05006959289312363,
0.06861566007137299,
0.0974634513258934,
-0.1729826033115387,
-0.057957619428634644,
0.1353495568037033,
-0.09663163125514984,
-0.14797286689281464,
0.10109983384609222,
-0.050693199038505554,
0.01928282529115677,
0.047161467373371124,
0.21422064304351807,
0.06321150809526443,
-0.09143578261137009,
-0.02580295503139496,
-0.03329068049788475,
0.04440530762076378,
-0.06494591385126114,
0.10175396502017975,
0.027680888772010803,
0.05365157872438431,
0.023984158411622047,
-0.032899804413318634,
0.03856229409575462,
-0.08387355506420135,
-0.10054522752761841,
-0.05070185661315918,
-0.08565592765808105,
0.039397966116666794,
0.05604296177625656,
0.05994046851992607,
-0.10856401175260544,
-0.09017187356948853,
0.04204317927360535,
0.0943065956234932,
-0.07395226508378983,
0.028962817043066025,
-0.09000826627016068,
0.11608705669641495,
-0.08325600624084473,
-0.02390553615987301,
-0.1791052222251892,
-0.04241684824228287,
0.040629271417856216,
-0.01629319041967392,
-0.006912850774824619,
-0.048891667276620865,
0.07074255496263504,
0.08783093094825745,
-0.05235742777585983,
-0.05203414335846901,
-0.05522594600915909,
0.008213330060243607,
-0.1105295866727829,
-0.1776295155286789,
-0.08015653491020203,
-0.0380605012178421,
0.14977632462978363,
-0.15268713235855103,
0.022225622087717056,
0.061116840690374374,
0.12500962615013123,
0.059797611087560654,
-0.04723487049341202,
-0.007436409126967192,
0.021452799439430237,
-0.05571167171001434,
-0.08678608387708664,
0.05719248577952385,
0.03528200834989548,
-0.07155010849237442,
-0.019102152436971664,
-0.10049699991941452,
0.1498662531375885,
0.13190734386444092,
-0.0015375686343759298,
-0.04512632265686989,
-0.01160994078963995,
-0.06610778719186783,
-0.030441991984844208,
-0.04081778973340988,
0.018804829567670822,
0.10142559558153152,
0.01744643971323967,
0.14419154822826385,
-0.09178037941455841,
-0.036961425095796585,
0.053544968366622925,
-0.028453968465328217,
-0.0331195667386055,
0.07361359149217606,
0.02190210297703743,
-0.14263916015625,
0.15015269815921783,
0.14882412552833557,
-0.04894813522696495,
0.12402692437171936,
-0.036747027188539505,
-0.0615357980132103,
-0.044876549392938614,
-0.037704430520534515,
0.014213677495718002,
0.1403394341468811,
-0.08333314955234528,
-0.005919712595641613,
0.05630137771368027,
0.019257593899965286,
-0.007085299585014582,
-0.18072617053985596,
0.0006808378966525197,
0.03521978110074997,
-0.04604950174689293,
-0.02278841845691204,
-0.014470276422798634,
0.0007941273506730795,
0.09172741323709488,
0.019804218783974648,
-0.07102026045322418,
0.05166372284293175,
0.010580740869045258,
-0.05623085796833038,
0.16415521502494812,
-0.07910753041505814,
-0.19727325439453125,
-0.11776646971702576,
-0.08754957467317581,
-0.10735819488763809,
0.013021474704146385,
0.06737184524536133,
-0.050448641180992126,
-0.04938974231481552,
-0.10206248611211777,
-0.04453543201088905,
0.021900271996855736,
0.02429220825433731,
0.05370878055691719,
-0.008031168952584267,
0.08405356109142303,
-0.09224440902471542,
-0.03291117399930954,
-0.014789600856602192,
0.018657125532627106,
0.06682770699262619,
0.018715238198637962,
0.11069032549858093,
0.08161229640245438,
-0.02844928205013275,
0.05646483600139618,
-0.01682325080037117,
0.2655041813850403,
-0.06765957176685333,
-0.006789656355977058,
0.13932959735393524,
-0.013368978165090084,
0.08428963273763657,
0.1268712729215622,
0.04151352122426033,
-0.09555158019065857,
-0.013173693791031837,
-0.00024822441628202796,
-0.05275752767920494,
-0.1537386178970337,
-0.04163756221532822,
-0.045641690492630005,
-0.0021682933438569307,
0.13930507004261017,
0.03818075731396675,
0.02474883571267128,
0.07807637751102448,
0.020041609182953835,
0.05664918199181557,
-0.017527885735034943,
0.10406769812107086,
0.08156019449234009,
0.06448414921760559,
0.13368317484855652,
-0.03653626888990402,
-0.019387291744351387,
0.05662747099995613,
0.04215037450194359,
0.20423758029937744,
-0.02541770040988922,
0.14701254665851593,
0.02641657367348671,
0.19307395815849304,
0.017521383240818977,
0.0728468969464302,
-0.014410126954317093,
0.0009393728105351329,
-0.019274147227406502,
-0.04702805355191231,
-0.06427313387393951,
0.03288881108164787,
-0.016649875789880753,
0.05632343888282776,
-0.09356046468019485,
0.039105307310819626,
0.059592608362436295,
0.30666422843933105,
0.06539998203516006,
-0.4122132360935211,
-0.09836560487747192,
0.012291035614907742,
0.0009865236934274435,
-0.055195607244968414,
-0.0072626820765435696,
0.0979013666510582,
-0.09949664771556854,
0.08215389400720596,
-0.09418605268001556,
0.08514873683452606,
-0.0845724418759346,
0.020298872143030167,
0.07689075917005539,
0.056060366332530975,
0.013226890936493874,
0.05964293330907822,
-0.21821673214435577,
0.24971400201320648,
0.018467964604496956,
0.04422129690647125,
-0.08908867090940475,
0.010060982778668404,
0.033364444971084595,
0.059161990880966187,
0.08554306626319885,
0.005977867171168327,
-0.09024009108543396,
-0.18880225718021393,
-0.1258762925863266,
0.0005427713040262461,
0.06169470399618149,
-0.036699384450912476,
0.09451829642057419,
-0.018175894394516945,
-0.012127134948968887,
0.021332256495952606,
0.0005201056483201683,
-0.03501477465033531,
-0.103630930185318,
0.02024604007601738,
0.034688886255025864,
-0.012138742953538895,
-0.06473075598478317,
-0.11475593596696854,
-0.03554871678352356,
0.16192500293254852,
0.05505121126770973,
-0.07524240761995316,
-0.1408705860376358,
0.07218684256076813,
0.07781627029180527,
-0.0855332687497139,
0.039305757731199265,
-0.016779718920588493,
0.14986851811408997,
0.020937321707606316,
-0.08943228423595428,
0.10178638249635696,
-0.05869165062904358,
-0.17860572040081024,
-0.041185978800058365,
0.09929849207401276,
0.007366738747805357,
0.05263189971446991,
0.004192214459180832,
0.06014186516404152,
-0.035002902150154114,
-0.0584394596517086,
0.06681792438030243,
-0.0073097143322229385,
0.10614755749702454,
-0.014883637428283691,
0.00864378735423088,
0.029195772483944893,
-0.04613848030567169,
0.00009839441918302327,
0.1684505194425583,
0.24079899489879608,
-0.10403203964233398,
0.060546230524778366,
0.03012177161872387,
-0.030879246070981026,
-0.18261685967445374,
0.010319743305444717,
0.07656802982091904,
-0.0001991603203350678,
0.04173794388771057,
-0.16060468554496765,
0.055176541209220886,
0.10514935851097107,
-0.043303944170475006,
0.08152011036872864,
-0.2768779397010803,
-0.11840421706438065,
0.0923023670911789,
0.138164222240448,
0.0691317543387413,
-0.13107311725616455,
-0.04327763617038727,
-0.041234806180000305,
-0.17335952818393707,
0.13665583729743958,
-0.05704028159379959,
0.11501350998878479,
-0.039327461272478104,
0.08051838725805283,
0.014901114627718925,
-0.056082114577293396,
0.14561402797698975,
0.005515002179890871,
0.08661133795976639,
-0.07185279577970505,
-0.0014093852369114757,
0.10643326491117477,
-0.10252601653337479,
0.07192501425743103,
-0.0869532898068428,
0.06187514215707779,
-0.10810889303684235,
-0.0037693935446441174,
-0.07425615191459656,
0.013987713493406773,
-0.013397954404354095,
-0.048907287418842316,
-0.0448833703994751,
0.03488645330071449,
0.06301422417163849,
-0.018155096098780632,
0.20988906919956207,
0.06445588916540146,
0.0862940326333046,
0.1728745847940445,
0.05397673323750496,
-0.10576145350933075,
-0.09408308565616608,
-0.04430058225989342,
-0.029343122616410255,
0.059755485504865646,
-0.13705183565616608,
0.053009506314992905,
0.12004052102565765,
0.013443393632769585,
0.1280696988105774,
0.05582417547702789,
-0.030783196911215782,
0.035687193274497986,
0.06206676363945007,
-0.1721130907535553,
-0.08640376478433609,
-0.010029762983322144,
0.030597826465964317,
-0.13003188371658325,
0.045725177973508835,
0.12137939780950546,
-0.0593545101583004,
-0.014887568540871143,
-0.004342919681221247,
0.03682979568839073,
-0.009421703405678272,
0.15946903824806213,
0.047883741557598114,
0.05509158596396446,
-0.11808934807777405,
0.11348052322864532,
0.057328153401613235,
-0.0728185623884201,
0.032391179352998734,
0.05030714347958565,
-0.10392948985099792,
-0.021465230733156204,
0.031419817358255386,
0.14932547509670258,
-0.06275127828121185,
-0.045640427619218826,
-0.13568063080310822,
-0.091814324259758,
0.06645428389310837,
0.07967224717140198,
0.0933644250035286,
0.01663324609398842,
-0.03539150580763817,
-0.013165266253054142,
-0.10855977237224579,
0.10982618480920792,
0.04324139654636383,
0.09105362743139267,
-0.17992232739925385,
0.054193608462810516,
-0.0015555275604128838,
0.07246194779872894,
-0.021836427971720695,
-0.00042325531831011176,
-0.08788467198610306,
0.003508437890559435,
-0.10813499987125397,
0.02464236691594124,
-0.052905477583408356,
0.006243168842047453,
-0.02064651995897293,
-0.0580705925822258,
-0.06364380568265915,
0.024784497916698456,
-0.11918067932128906,
-0.053243763744831085,
0.02146504819393158,
0.031834639608860016,
-0.12016978859901428,
-0.04392008110880852,
0.020345089957118034,
-0.08986733108758926,
0.09774119406938553,
0.06029992923140526,
-0.008077923208475113,
0.00773270707577467,
0.0036002967972308397,
-0.02274298295378685,
0.0666942149400711,
0.007561622653156519,
0.08597277849912643,
-0.1152612566947937,
-0.0221384409815073,
0.01634843461215496,
-0.004547150805592537,
0.017726117745041847,
0.15840598940849304,
-0.12086156010627747,
-0.0003179961640853435,
-0.014678256586194038,
-0.06600851565599442,
-0.06344839930534363,
0.06893838196992874,
0.10903503000736237,
0.02346671372652054,
0.21181334555149078,
-0.054371658712625504,
0.015811823308467865,
-0.20995409786701202,
-0.011581460013985634,
0.005185890011489391,
-0.1388559192419052,
-0.10497695952653885,
-0.03237957879900932,
0.06376256048679352,
-0.07031478732824326,
0.11765085160732269,
0.03525954857468605,
0.02161695808172226,
0.02906344085931778,
0.025029366835951805,
-0.0031726681627333164,
0.013450034894049168,
0.16309522092342377,
0.014403261244297028,
-0.028442582115530968,
0.12852592766284943,
0.028986822813749313,
0.09334488213062286,
0.11778779327869415,
0.17672526836395264,
0.11388354748487473,
0.04729508236050606,
0.09055530279874802,
0.05202596262097359,
-0.025968270376324654,
-0.22174733877182007,
0.03601896017789841,
-0.03978736698627472,
0.1488790065050125,
-0.0030294209718704224,
0.15902450680732727,
0.0920415073633194,
-0.18360793590545654,
0.040488436818122864,
-0.03700747340917587,
-0.0790853351354599,
-0.08454839885234833,
-0.12155362963676453,
-0.10311590880155563,
-0.15089921653270721,
0.002945262473076582,
-0.1040843203663826,
0.023338600993156433,
0.11202728003263474,
-0.008582104928791523,
-0.009919910691678524,
0.116677425801754,
-0.02631515823304653,
0.026041926816105843,
0.03836518153548241,
0.00608045794069767,
-0.059937771409749985,
-0.044151950627565384,
-0.08065995573997498,
0.014101422391831875,
0.032313644886016846,
0.05599058046936989,
-0.03235676884651184,
-0.007023791316896677,
0.03841041401028633,
-0.010091220960021019,
-0.12353866547346115,
0.01347822230309248,
0.005028906278312206,
0.05164548382163048,
0.0008541525457985699,
0.012780209071934223,
0.03201600909233093,
-0.015217483974993229,
0.19341084361076355,
-0.07325411587953568,
-0.027416478842496872,
-0.1228807121515274,
0.17896701395511627,
0.0026140701957046986,
-0.04994320869445801,
0.05295133590698242,
-0.09137362241744995,
-0.020702529698610306,
0.15485265851020813,
0.1892986297607422,
-0.07158271223306656,
-0.016520513221621513,
-0.017527583986520767,
-0.013897030614316463,
-0.022615507245063782,
0.09919055551290512,
0.0991419330239296,
-0.0069245584309101105,
-0.0751221776008606,
-0.028980256989598274,
-0.06606413424015045,
-0.034512959420681,
-0.03850788250565529,
0.06925404816865921,
-0.004570751916617155,
0.0070457919500768185,
-0.07483471930027008,
0.04310325160622597,
-0.02210995741188526,
-0.06085818260908127,
0.06226903945207596,
-0.21256737411022186,
-0.17790570855140686,
0.006773421075195074,
0.07538973540067673,
0.0015973751433193684,
0.0461571104824543,
-0.009913075715303421,
0.018662674352526665,
0.07594356685876846,
-0.02225665934383869,
-0.08672447502613068,
-0.09593749046325684,
0.10812120139598846,
-0.13375911116600037,
0.2528570294380188,
-0.03883460536599159,
0.03583916276693344,
0.12127543240785599,
0.041867125779390335,
-0.1335451751947403,
0.03351692110300064,
0.03981999680399895,
-0.032485269010066986,
0.00548918079584837,
0.14240407943725586,
-0.03740047290921211,
0.07958021014928818,
0.0458458811044693,
-0.1027912050485611,
-0.03964604437351227,
-0.04966754838824272,
-0.011354409158229828,
-0.024445757269859314,
-0.054610975086688995,
-0.036348532885313034,
0.13227923214435577,
0.17175258696079254,
-0.042096637189388275,
-0.023690558969974518,
-0.06475082784891129,
0.030860183760523796,
0.07729368656873703,
-0.03295742720365524,
-0.052064236253499985,
-0.23603148758411407,
0.0024359924718737602,
0.05229694023728371,
-0.013576737605035305,
-0.20701472461223602,
-0.110505111515522,
0.0060418094508349895,
-0.05801977962255478,
-0.07628542929887772,
0.09231390058994293,
0.06255589425563812,
0.035103797912597656,
-0.06320928037166595,
0.038133736699819565,
-0.07872021943330765,
0.14179112017154694,
-0.14508864283561707,
-0.07859515398740768
] |
null | null | transformers |
# Model Card for Mistral-7B-Instruct-v0.1
The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) generative text model using a variety of publicly available conversation datasets.
For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/announcing-mistral-7b/).
## Instruction format
In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
E.g.
```
text = "<s>[INST] What is your favourite condiment? [/INST]"
"Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
"[INST] Do you have mayonnaise recipes? [/INST]"
```
This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
device = "cuda" # the device to load the model onto
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
messages = [
{"role": "user", "content": "What is your favourite condiment?"},
{"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"},
{"role": "user", "content": "Do you have mayonnaise recipes?"}
]
encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
model_inputs = encodeds.to(device)
model.to(device)
generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
```
## Model Architecture
This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:
- Grouped-Query Attention
- Sliding-Window Attention
- Byte-fallback BPE tokenizer
## Troubleshooting
- If you see the following error:
```
Traceback (most recent call last):
File "", line 1, in
File "/transformers/models/auto/auto_factory.py", line 482, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
File "/transformers/models/auto/configuration_auto.py", line 1022, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
File "/transformers/models/auto/configuration_auto.py", line 723, in getitem
raise KeyError(key)
KeyError: 'mistral'
```
Installing transformers from source should solve the issue
pip install git+https://github.com/huggingface/transformers
This should not be required after transformers-v4.33.4.
## Limitations
The Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance.
It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
## The Mistral AI Team
Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed. | {"license": "apache-2.0", "tags": ["finetuned"], "pipeline_tag": "text-generation", "inference": false} | text-generation | gadkins/Mistral-7B-Instruct-v0.1-function-calling | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"conversational",
"arxiv:2310.06825",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:21:55+00:00 | [
"2310.06825"
] | [] | TAGS
#transformers #safetensors #mistral #text-generation #finetuned #conversational #arxiv-2310.06825 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us
|
# Model Card for Mistral-7B-Instruct-v0.1
The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.1 generative text model using a variety of publicly available conversation datasets.
For full details of this model please read our paper and release blog post.
## Instruction format
In order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
E.g.
This format is available as a chat template via the 'apply_chat_template()' method:
## Model Architecture
This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:
- Grouped-Query Attention
- Sliding-Window Attention
- Byte-fallback BPE tokenizer
## Troubleshooting
- If you see the following error:
Installing transformers from source should solve the issue
pip install git+URL
This should not be required after transformers-v4.33.4.
## Limitations
The Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance.
It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
## The Mistral AI Team
Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed. | [
"# Model Card for Mistral-7B-Instruct-v0.1\n\nThe Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.1 generative text model using a variety of publicly available conversation datasets.\n\nFor full details of this model please read our paper and release blog post.",
"## Instruction format\n\nIn order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.\n\nE.g.\n\n\nThis format is available as a chat template via the 'apply_chat_template()' method:",
"## Model Architecture\nThis instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:\n- Grouped-Query Attention\n- Sliding-Window Attention\n- Byte-fallback BPE tokenizer",
"## Troubleshooting\n- If you see the following error:\n\n\nInstalling transformers from source should solve the issue\npip install git+URL\n\nThis should not be required after transformers-v4.33.4.",
"## Limitations\n\nThe Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance. \nIt does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to\nmake the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.",
"## The Mistral AI Team\n\nAlbert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed."
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #finetuned #conversational #arxiv-2310.06825 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us \n",
"# Model Card for Mistral-7B-Instruct-v0.1\n\nThe Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.1 generative text model using a variety of publicly available conversation datasets.\n\nFor full details of this model please read our paper and release blog post.",
"## Instruction format\n\nIn order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.\n\nE.g.\n\n\nThis format is available as a chat template via the 'apply_chat_template()' method:",
"## Model Architecture\nThis instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:\n- Grouped-Query Attention\n- Sliding-Window Attention\n- Byte-fallback BPE tokenizer",
"## Troubleshooting\n- If you see the following error:\n\n\nInstalling transformers from source should solve the issue\npip install git+URL\n\nThis should not be required after transformers-v4.33.4.",
"## Limitations\n\nThe Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance. \nIt does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to\nmake the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.",
"## The Mistral AI Team\n\nAlbert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed."
] | [
64,
81,
105,
56,
42,
85,
100
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #finetuned #conversational #arxiv-2310.06825 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us \n# Model Card for Mistral-7B-Instruct-v0.1\n\nThe Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.1 generative text model using a variety of publicly available conversation datasets.\n\nFor full details of this model please read our paper and release blog post.## Instruction format\n\nIn order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.\n\nE.g.\n\n\nThis format is available as a chat template via the 'apply_chat_template()' method:## Model Architecture\nThis instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:\n- Grouped-Query Attention\n- Sliding-Window Attention\n- Byte-fallback BPE tokenizer## Troubleshooting\n- If you see the following error:\n\n\nInstalling transformers from source should solve the issue\npip install git+URL\n\nThis should not be required after transformers-v4.33.4.## Limitations\n\nThe Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance. \nIt does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to\nmake the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs."
] | [
-0.08154156804084778,
0.09633228182792664,
-0.006207073573023081,
0.008780199103057384,
0.1072206050157547,
-0.007847028784453869,
0.10169675201177597,
0.06143496558070183,
0.031456686556339264,
0.05882946029305458,
-0.00011387487757019699,
0.07851465046405792,
0.06145763024687767,
0.12915445864200592,
0.008660025894641876,
-0.19192884862422943,
0.0663599744439125,
-0.05721339210867882,
0.08362677693367004,
0.06200484558939934,
0.1273060292005539,
-0.04884813725948334,
0.03294702246785164,
0.0712774395942688,
-0.08009567111730576,
-0.0010871957056224346,
-0.0017357796896249056,
-0.020775537937879562,
0.08311796933412552,
0.061596859246492386,
0.04196586459875107,
0.021883295848965645,
0.006327025592327118,
-0.1697111576795578,
0.017486199736595154,
0.08615092933177948,
-0.02327769808471203,
0.05837631970643997,
0.08899974077939987,
0.004212993197143078,
0.0772939994931221,
-0.09041368961334229,
-0.009055795148015022,
0.058675553649663925,
-0.04143316298723221,
-0.1454366147518158,
-0.07103773951530457,
0.11868322640657425,
0.12445646524429321,
0.07823767513036728,
-0.016289163380861282,
0.04367344081401825,
0.0395798534154892,
0.09049387276172638,
0.14002874493598938,
-0.19876842200756073,
-0.023655302822589874,
0.01687064953148365,
0.05342654511332512,
0.06372920423746109,
-0.05198855325579643,
0.0009066116763278842,
0.005140439607203007,
0.021791601553559303,
0.024018293246626854,
-0.0019940088968724012,
0.0009535275166854262,
-0.055480606853961945,
-0.12282145023345947,
-0.030890386551618576,
0.19882136583328247,
0.014046670868992805,
-0.06924184411764145,
-0.1450638324022293,
-0.08223876357078552,
0.10178078711032867,
0.0464634895324707,
-0.060642555356025696,
0.04725273326039314,
0.044318344444036484,
0.11003292351961136,
-0.10921680927276611,
-0.1320272833108902,
-0.021872414276003838,
-0.03310839831829071,
0.06860223412513733,
0.005207359790802002,
0.042009320110082626,
-0.052894845604896545,
0.09600532799959183,
-0.08062339574098587,
-0.07835927605628967,
-0.09532977640628815,
-0.005415583960711956,
-0.0651906430721283,
-0.05199264734983444,
-0.01091353315860033,
-0.12757200002670288,
0.034489139914512634,
0.1836443841457367,
-0.11042192578315735,
0.06038421764969826,
0.049516040831804276,
0.023722048848867416,
0.03355678170919418,
0.12768621742725372,
-0.03221641108393669,
-0.019580919295549393,
0.0768846794962883,
0.02044259011745453,
0.10814391821622849,
0.013155734166502953,
-0.053735122084617615,
-0.07165951281785965,
0.06459255516529083,
0.04131361097097397,
0.04869859665632248,
0.06227762997150421,
-0.0376102551817894,
-0.001306664664298296,
0.26516038179397583,
-0.0967763364315033,
0.02814355678856373,
0.013017978519201279,
-0.02028176374733448,
0.033470384776592255,
0.07905180007219315,
-0.022962583228945732,
-0.07819421589374542,
-0.024468882009387016,
-0.06177689507603645,
-0.03168104216456413,
-0.06586197763681412,
-0.09287981688976288,
0.021313564851880074,
-0.022859958931803703,
-0.03926181420683861,
-0.12873655557632446,
-0.23771782219409943,
-0.01534868124872446,
0.01893850974738598,
-0.018686464056372643,
-0.001529863802716136,
-0.03305979073047638,
-0.06259134411811829,
-0.0023101656697690487,
-0.03646128624677658,
0.0023093034978955984,
-0.03734690323472023,
0.007707825396209955,
-0.0059781321324408054,
0.0298141036182642,
-0.044473569840192795,
0.002339709084481001,
-0.043771978467702866,
0.03075489029288292,
-0.21934865415096283,
0.0963660329580307,
-0.009434050880372524,
0.02644040621817112,
-0.09150896221399307,
0.004362340085208416,
0.08654904365539551,
0.01604987308382988,
0.05475911498069763,
0.12660352885723114,
-0.12970206141471863,
0.009858810342848301,
0.11105254292488098,
-0.1726243644952774,
-0.02429627627134323,
0.1350969821214676,
0.03674058988690376,
0.017298176884651184,
0.08462117612361908,
0.0882025733590126,
0.0671641081571579,
-0.08341231197118759,
-0.08670376241207123,
0.042187828570604324,
-0.10596194118261337,
-0.002574948128312826,
0.008858786895871162,
-0.0020801806822419167,
0.06598709523677826,
0.024585096165537834,
-0.061075396835803986,
0.04520360007882118,
0.010262117721140385,
0.03124273009598255,
-0.012454826384782791,
-0.05076809227466583,
-0.010683098807930946,
-0.031673677265644073,
-0.026618700474500656,
-0.04481608793139458,
-0.11877778172492981,
0.08299194276332855,
0.12774930894374847,
-0.016659533604979515,
0.01906248740851879,
-0.10034265369176865,
0.11269867420196533,
-0.045037735253572464,
0.005332382395863533,
-0.14447365701198578,
-0.07569418847560883,
0.03399287536740303,
-0.02604662999510765,
-0.039023250341415405,
0.050965677946805954,
0.042302753776311874,
0.04952188581228256,
0.026010168716311455,
-0.06521331518888474,
0.06770047545433044,
-0.014859385788440704,
-0.09698063135147095,
-0.08673080801963806,
-0.033755771815776825,
-0.05574924871325493,
0.14501607418060303,
-0.1826154589653015,
0.061888810247182846,
0.10896552354097366,
0.12557123601436615,
0.0024079422000795603,
-0.05644238740205765,
0.023873742669820786,
-0.07088005542755127,
-0.030740030109882355,
-0.06980383396148682,
0.03163948655128479,
0.06980279833078384,
-0.0011772493598982692,
0.07082220166921616,
-0.1891135275363922,
-0.14061352610588074,
0.038038887083530426,
0.03334725648164749,
-0.04045233130455017,
-0.05103394389152527,
-0.024245675653219223,
-0.027028605341911316,
-0.04528401046991348,
-0.14133207499980927,
0.19857902824878693,
0.004294462967664003,
0.11219455301761627,
-0.059291303157806396,
-0.08822697401046753,
-0.019659575074911118,
-0.03592917323112488,
-0.06327211856842041,
0.05531343072652817,
-0.14303137362003326,
-0.12742403149604797,
0.05741044133901596,
-0.00451402785256505,
-0.06315154582262039,
0.07461755722761154,
0.035855576395988464,
-0.020374873653054237,
-0.0031368499621748924,
0.08604368567466736,
0.012112190015614033,
0.032661497592926025,
-0.085082046687603,
-0.03065667673945427,
0.027008987963199615,
0.033963434398174286,
0.04762523993849754,
-0.10410184413194656,
0.08678242564201355,
0.021484499797225,
-0.035339709371328354,
0.042823344469070435,
0.01825811341404915,
-0.03719064220786095,
0.05070733651518822,
-0.007173601072281599,
0.02133331261575222,
0.039273701608181,
-0.056529149413108826,
-0.13520567119121552,
0.15119893848896027,
-0.1058342307806015,
-0.18273593485355377,
-0.1465224325656891,
-0.05167856067419052,
-0.03484993800520897,
0.012683078646659851,
0.08053985238075256,
-0.05845479294657707,
-0.09086712449789047,
-0.09263724833726883,
0.007674391847103834,
-0.004250910133123398,
-0.06306520849466324,
0.012293579056859016,
-0.06363850086927414,
0.052809491753578186,
-0.13972529768943787,
0.0024721070658415556,
0.016013385728001595,
-0.08422306925058365,
0.017413288354873657,
0.014346539974212646,
0.04018651321530342,
0.08642205595970154,
-0.0259703379124403,
-0.009339735843241215,
-0.002734991954639554,
0.20375201106071472,
-0.027849530801177025,
0.131210595369339,
0.2488357275724411,
-0.014493550173938274,
0.09078287333250046,
0.08262913674116135,
-0.04555415362119675,
-0.03818519786000252,
0.024888355284929276,
-0.01268477737903595,
-0.007882056757807732,
-0.20908638834953308,
-0.033334385603666306,
-0.03944312036037445,
-0.06634820997714996,
0.01081067230552435,
0.07721099257469177,
0.08383850008249283,
0.06279320269823074,
-0.05938812345266342,
0.02519209124147892,
0.0861630067229271,
0.11334945261478424,
0.10871215164661407,
0.006266547832638025,
0.02595028653740883,
-0.05080956965684891,
0.06583000719547272,
0.09765420854091644,
0.044645413756370544,
0.1280670166015625,
-0.09888101369142532,
0.19805961847305298,
0.04413682594895363,
0.002376264426857233,
0.019132398068904877,
0.06683413684368134,
-0.08785507082939148,
0.017204483970999718,
-0.005850804969668388,
-0.06235991045832634,
-0.007531102746725082,
0.0798039361834526,
-0.0816178172826767,
0.05089205875992775,
-0.05062584578990936,
0.04774991795420647,
0.07080574333667755,
0.21534864604473114,
0.02865188755095005,
-0.19363798201084137,
-0.10511902719736099,
0.0749790221452713,
-0.05346699431538582,
-0.10758927464485168,
-0.009356308728456497,
0.1667044460773468,
-0.05342321842908859,
0.0306096188724041,
-0.008031410165131092,
0.07804418355226517,
-0.11466698348522186,
-0.00043245372944511473,
-0.008635341189801693,
0.24105298519134521,
-0.008597852662205696,
0.07308010011911392,
-0.10082577168941498,
0.052874550223350525,
0.00893736444413662,
0.11089503765106201,
-0.07552731037139893,
0.07359142601490021,
0.028938282281160355,
0.08730493485927582,
0.11816561967134476,
0.0003166341921314597,
-0.02291649393737316,
0.00797887984663248,
-0.048014357686042786,
0.0014339331537485123,
0.01018709409981966,
-0.030359379947185516,
0.03409074246883392,
-0.03731171786785126,
-0.020964914932847023,
0.0068831066600978374,
-0.008402328006923199,
-0.12321994453668594,
-0.1317293494939804,
0.00708656245842576,
0.09863388538360596,
0.04076865315437317,
-0.08192809671163559,
-0.011751317419111729,
-0.10663830488920212,
0.1972830593585968,
-0.10674576461315155,
-0.1261521875858307,
-0.11902682483196259,
-0.02137533389031887,
0.03745195269584656,
-0.05910748988389969,
0.027378607541322708,
-0.017924176529049873,
0.15250934660434723,
-0.026262616738677025,
-0.045752622187137604,
0.023749997839331627,
-0.13060764968395233,
-0.09092611074447632,
-0.02605222910642624,
0.028365764766931534,
0.08891803026199341,
-0.01916622929275036,
0.016962802037596703,
-0.017019912600517273,
-0.03530968353152275,
-0.07609854638576508,
-0.009652198292315006,
0.21714167296886444,
0.07045773416757584,
0.03359309211373329,
-0.007985609583556652,
-0.18548142910003662,
-0.030345916748046875,
-0.03672707453370094,
0.048331599682569504,
0.2562737762928009,
-0.049217820167541504,
0.05058899149298668,
0.18936152756214142,
-0.06738542020320892,
-0.15720102190971375,
0.007740884553641081,
0.025884414091706276,
0.033461783081293106,
-0.02465093694627285,
-0.0896829217672348,
0.11012634634971619,
0.07694191485643387,
-0.01870664581656456,
0.07745762914419174,
-0.16128404438495636,
-0.09880801290273666,
0.042679525911808014,
0.09216197580099106,
0.07744354754686356,
-0.07940661162137985,
-0.05761032924056053,
-0.026216130703687668,
-0.14368797838687897,
-0.02632373385131359,
-0.055871784687042236,
0.03519465774297714,
-0.0026134061627089977,
0.018893739208579063,
0.031505610793828964,
-0.06515925377607346,
0.10333551466464996,
-0.018474703654646873,
0.05539817735552788,
-0.08083850145339966,
0.026569433510303497,
0.053962815552949905,
-0.09549470245838165,
0.1264963299036026,
-0.09966200590133667,
0.05628080666065216,
-0.030570123344659805,
-0.03545195236802101,
-0.07858866453170776,
0.06015896797180176,
-0.015529200434684753,
-0.0515558235347271,
-0.011435618624091148,
0.019585909321904182,
0.0483914315700531,
0.007322723977267742,
0.0026524229906499386,
-0.05895558372139931,
0.007541017606854439,
0.13097883760929108,
0.15177416801452637,
-0.04881967231631279,
-0.0518326573073864,
-0.02068394236266613,
-0.018698979169130325,
0.05299860239028931,
-0.06150960549712181,
0.012772730551660061,
0.04915092885494232,
0.008852294646203518,
0.13488848507404327,
0.027577048167586327,
-0.12567901611328125,
0.014431819319725037,
0.038031116127967834,
-0.08139865100383759,
-0.13873031735420227,
-0.04467925801873207,
0.1593923717737198,
-0.06571009010076523,
0.07613122463226318,
0.13893690705299377,
-0.018571294844150543,
-0.023382579907774925,
0.005920398980379105,
0.03915601223707199,
-0.01629496179521084,
0.04275176674127579,
-0.03086734563112259,
0.05831968039274216,
-0.03304346278309822,
0.07501194626092911,
0.061046987771987915,
-0.061243683099746704,
0.027796011418104172,
0.13632553815841675,
-0.12669850885868073,
-0.06473308056592941,
-0.08565132319927216,
0.04261361435055733,
-0.009116536006331444,
-0.08158857375383377,
-0.024359362199902534,
-0.06662381440401077,
0.0024263656232506037,
0.1134682297706604,
0.022950429469347,
-0.017558936029672623,
0.0065427254885435104,
0.04917560890316963,
-0.09629447758197784,
0.08463370054960251,
0.0029047136195003986,
0.03889460861682892,
-0.12658731639385223,
0.047439221292734146,
0.007650384679436684,
-0.006139704491943121,
-0.0028994774911552668,
-0.02040470764040947,
-0.09330154210329056,
-0.02214830368757248,
-0.17315718531608582,
0.04985164850950241,
-0.11203234642744064,
-0.008426605723798275,
0.016963593661785126,
-0.015599031932651997,
0.011325571686029434,
0.08314673602581024,
0.000001967358684851206,
-0.03224293142557144,
-0.0039048076141625643,
0.06008775159716606,
-0.11691708117723465,
-0.037147156894207,
-0.040404096245765686,
-0.09426525980234146,
0.15988989174365997,
0.015690991654992104,
-0.044613949954509735,
-0.03224821761250496,
-0.2084404081106186,
0.0007246200111694634,
0.01057891920208931,
0.009058808907866478,
0.025258924812078476,
-0.10466037690639496,
-0.006423517595976591,
0.014673789031803608,
-0.07853790372610092,
-0.05231739580631256,
0.04035340994596481,
-0.029848041012883186,
0.06629069149494171,
0.004589376505464315,
0.005064367316663265,
-0.1202789694070816,
0.0279405377805233,
0.12313489615917206,
0.017476752400398254,
0.09651067852973938,
-0.06591952592134476,
0.0042338902130723,
-0.09558138996362686,
0.003855651244521141,
0.05433174595236778,
-0.025013215839862823,
-0.028623688966035843,
-0.002451494336128235,
0.036656420677900314,
-0.03735818713903427,
0.02209842950105667,
-0.056990694254636765,
-0.01916835457086563,
0.04664567857980728,
0.011928623542189598,
-0.05829600617289543,
-0.00361150735989213,
0.053269125521183014,
-0.057679302990436554,
0.023410018533468246,
0.006617214065045118,
0.006312564946711063,
0.044750768691301346,
-0.029643243178725243,
0.1229892298579216,
0.07399932295084,
-0.017640559002757072,
0.13844211399555206,
-0.01697414740920067,
-0.06044105812907219,
-0.16914434731006622,
0.02278098650276661,
-0.07075510174036026,
0.05179936811327934,
-0.08297117799520493,
0.04691582918167114,
0.20129500329494476,
-0.07798559218645096,
0.011849405243992805,
-0.01771857589483261,
-0.036326996982097626,
-0.05591390281915665,
-0.2381640523672104,
-0.016117876395583153,
-0.047758180648088455,
-0.03483477979898453,
-0.09046582877635956,
0.03890863060951233,
0.030407220125198364,
0.00990974996238947,
0.015606806613504887,
0.062398962676525116,
0.01119961403310299,
-0.03606052324175835,
0.024931803345680237,
-0.0452849343419075,
0.046057019382715225,
0.018452554941177368,
0.0030803850386291742,
0.07534752041101456,
0.029167676344513893,
0.05953400954604149,
0.07267474383115768,
0.10500656068325043,
0.03795269504189491,
0.0025216196663677692,
-0.07427169382572174,
-0.009539584629237652,
0.0560164749622345,
-0.030736170709133148,
0.10405004769563675,
0.06410369277000427,
-0.03006855584681034,
0.004622407723218203,
0.1335315704345703,
-0.036748841404914856,
-0.09392099827528,
-0.10331721603870392,
0.1849967986345291,
-0.03214326500892639,
-0.01512314286082983,
0.03380288928747177,
-0.15556326508522034,
0.025377217680215836,
0.1749695986509323,
0.04319421574473381,
0.007487283553928137,
0.0009118511225096881,
-0.03164656087756157,
-0.014528177678585052,
-0.030137691646814346,
0.04467734694480896,
0.05967135354876518,
0.2726263105869293,
-0.019725125283002853,
0.10422668606042862,
0.014394399709999561,
-0.001594762783497572,
-0.04495081678032875,
0.03122458979487419,
-0.06876815855503082,
0.02526176907122135,
0.006550353020429611,
0.02376154251396656,
-0.017244193702936172,
-0.1525002419948578,
-0.03796863928437233,
0.03824366256594658,
-0.051701292395591736,
0.01886490173637867,
0.08170425146818161,
-0.025552472099661827,
0.07091312110424042,
-0.005825628060847521,
-0.014529180712997913,
0.2089289277791977,
-0.05452161282300949,
-0.0697922632098198,
0.007455307990312576,
0.0324506051838398,
-0.04078691452741623,
0.17204244434833527,
-0.01238649245351553,
0.0774761363863945,
0.11342265456914902,
0.017118770629167557,
-0.18192410469055176,
0.04801958054304123,
-0.034536682069301605,
-0.12090425193309784,
0.03440277650952339,
0.11000712215900421,
-0.03254379704594612,
0.056714147329330444,
0.024429315701127052,
-0.11966989189386368,
0.01585901528596878,
0.07929470390081406,
-0.01261954102665186,
-0.09587472677230835,
0.030387116596102715,
-0.06513641774654388,
0.14087893068790436,
0.05294308438897133,
-0.016314610838890076,
-0.03327777236700058,
-0.02926018089056015,
0.021609893068671227,
0.012252978049218655,
0.10078724473714828,
0.006912493146955967,
-0.14428693056106567,
0.05386963486671448,
0.011027777567505836,
0.06244484707713127,
-0.1579630970954895,
-0.0618620291352272,
-0.06199503317475319,
-0.05127492547035217,
-0.09001004695892334,
0.041248027235269547,
0.1338510662317276,
0.06048567593097687,
-0.029298514127731323,
0.047077205032110214,
-0.03153504058718681,
0.1118692010641098,
-0.09002561867237091,
-0.08042320609092712
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - Samoi/mimi
<Gallery />
## Model description
These are Samoi/mimi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using [DreamBooth](https://dreambooth.github.io/).
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of mimimi cat to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](Samoi/mimi/tree/main) them in the Files & versions tab.
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "widget": [{"text": "A photo of mimimi cat in a bucket", "output": {"url": "image_0.png"}}, {"text": "A photo of mimimi cat in a bucket", "output": {"url": "image_1.png"}}, {"text": "A photo of mimimi cat in a bucket", "output": {"url": "image_2.png"}}, {"text": "A photo of mimimi cat in a bucket", "output": {"url": "image_3.png"}}], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "a photo of mimimi cat"} | text-to-image | Samoi/mimi | [
"diffusers",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-06T06:24:59+00:00 | [] | [] | TAGS
#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - Samoi/mimi
<Gallery />
## Model description
These are Samoi/mimi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of mimimi cat to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
| [
"# SDXL LoRA DreamBooth - Samoi/mimi\n\n<Gallery />",
"## Model description\n\nThese are Samoi/mimi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of mimimi cat to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
"TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - Samoi/mimi\n\n<Gallery />",
"## Model description\n\nThese are Samoi/mimi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of mimimi cat to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
82,
19,
84,
19,
28
] | [
"passage: TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - Samoi/mimi\n\n<Gallery />## Model description\n\nThese are Samoi/mimi LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.## Trigger words\n\nYou should use a photo of mimimi cat to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
-0.07323776930570602,
0.1373778134584427,
0.0008950817282311618,
0.03581885248422623,
0.16353446245193481,
0.03289283439517021,
0.15238776803016663,
0.07972008734941483,
0.04729083925485611,
0.05916251987218857,
0.009752539917826653,
0.022786438465118408,
0.05051017925143242,
0.20628586411476135,
-0.0341084785759449,
-0.17340077459812164,
0.030142785981297493,
-0.012765240855515003,
-0.026404032483696938,
0.004539403133094311,
0.030575241893529892,
-0.07424163073301315,
0.07119353115558624,
-0.051844868808984756,
-0.07180151343345642,
0.04338885098695755,
-0.004896384198218584,
-0.022190440446138382,
0.010161012411117554,
0.046881236135959625,
0.0875430479645729,
0.04974127933382988,
0.08372192084789276,
-0.15682704746723175,
0.03501250967383385,
0.08155472576618195,
-0.024743424728512764,
0.022489165887236595,
-0.034062035381793976,
-0.07367642223834991,
0.08522354066371918,
-0.1024731993675232,
0.03524920716881752,
0.05854023993015289,
-0.008587370626628399,
-0.22787119448184967,
0.009758265689015388,
-0.05765021964907646,
0.06345739960670471,
0.06798326224088669,
0.005526803433895111,
0.039251659065485,
0.027928926050662994,
0.06233499199151993,
0.21701596677303314,
-0.11812654882669449,
-0.05604754760861397,
0.2253051996231079,
-0.002563607646152377,
0.15960095822811127,
-0.02014271169900894,
0.06718061864376068,
0.11913064867258072,
-0.0085061714053154,
0.08717595040798187,
-0.0594211146235466,
-0.030001452192664146,
-0.09325356036424637,
-0.09834953397512436,
0.05159124732017517,
0.23378576338291168,
-0.028760652989149094,
-0.09787964820861816,
-0.12083566933870316,
-0.056199826300144196,
0.1226724311709404,
-0.03537149354815483,
0.027280233800411224,
0.0316011980175972,
0.008198103867471218,
0.0487971194088459,
-0.09300891309976578,
-0.04236368462443352,
-0.10940547287464142,
0.05127901956439018,
0.19577690958976746,
0.010292916558682919,
0.027922315523028374,
0.011146442033350468,
0.08513122797012329,
-0.12133264541625977,
-0.13916747272014618,
-0.021699972450733185,
-0.03651595860719681,
0.010071401484310627,
0.04914511367678642,
-0.010860472917556763,
-0.16023840010166168,
0.05221039056777954,
-0.004165194928646088,
0.0067940643057227135,
0.019987883046269417,
-0.0634358823299408,
0.03340921178460121,
-0.03764262795448303,
0.07070250809192657,
0.002659481018781662,
-0.07501475512981415,
0.06989620625972748,
0.07037755101919174,
0.10016439110040665,
-0.03834925591945648,
-0.13408349454402924,
-0.015718374401330948,
-0.1320473998785019,
0.0911741778254509,
-0.022604405879974365,
-0.018850261345505714,
-0.08196180313825607,
-0.03395245224237442,
0.11890947818756104,
-0.06701382249593735,
0.006118092220276594,
-0.0360429622232914,
-0.04201781377196312,
0.09277667105197906,
0.14493420720100403,
-0.016701234504580498,
-0.000696960894856602,
0.05904438719153404,
-0.05096697062253952,
0.033855535089969635,
-0.09553781896829605,
-0.13318142294883728,
-0.0034545809030532837,
-0.1375761777162552,
0.00980337057262659,
-0.13609935343265533,
-0.24391956627368927,
-0.03025597147643566,
0.02165398746728897,
-0.012746592983603477,
0.07906508445739746,
-0.06839338690042496,
-0.07191821187734604,
-0.025184180587530136,
0.05089006572961807,
-0.0035516899079084396,
-0.018241042271256447,
0.04516197368502617,
-0.010404547676444054,
0.11998038738965988,
-0.06521693617105484,
0.025517547503113747,
-0.08754628896713257,
0.03239230066537857,
-0.12531912326812744,
0.11182178556919098,
-0.10594218224287033,
0.06733249872922897,
-0.053021255880594254,
-0.05191242694854736,
-0.03334977850317955,
0.0073915203101933,
0.021530045196413994,
0.11238417029380798,
-0.26794901490211487,
-0.027758589014410973,
0.1612067073583603,
-0.22135476768016815,
-0.07539907842874527,
0.06445500254631042,
-0.007778974715620279,
0.1077912300825119,
0.09627962857484818,
0.13235366344451904,
0.1963878571987152,
-0.21178406476974487,
-0.02408098243176937,
0.006068329326808453,
0.033598773181438446,
-0.053099125623703,
0.056836675852537155,
0.045478787273168564,
-0.05450478196144104,
0.059233345091342926,
-0.13619855046272278,
0.07674192637205124,
0.013380146585404873,
-0.029100704938173294,
-0.03904179483652115,
-0.1528904139995575,
0.015202292241156101,
0.029002970084547997,
0.017591549083590508,
-0.003129744902253151,
-0.009774132631719112,
0.07821384072303772,
0.15356376767158508,
-0.09905777126550674,
-0.006216211710125208,
0.009170383214950562,
0.10804244130849838,
-0.11595543473958969,
-0.005930280312895775,
-0.09316473454236984,
-0.08305290341377258,
0.05867480859160423,
0.0505332425236702,
0.013393462635576725,
0.004059405066072941,
0.05386862903833389,
0.11399766802787781,
-0.06281322240829468,
-0.06223587691783905,
0.030600009486079216,
-0.003911281004548073,
-0.005714581813663244,
-0.0965103879570961,
0.0009177628671750426,
-0.09183239191770554,
0.08534853905439377,
-0.1879771649837494,
0.07287982106208801,
-0.012720469385385513,
0.09641151875257492,
0.10916934907436371,
-0.03866991773247719,
0.0709272027015686,
0.0361148864030838,
-0.03261987119913101,
-0.09176509827375412,
-0.013813900761306286,
-0.0002594370162114501,
-0.153212770819664,
0.07978826016187668,
-0.17846210300922394,
0.055300522595644,
0.11025496572256088,
0.14923810958862305,
-0.018810158595442772,
-0.05668768286705017,
0.03648859262466431,
0.009771917015314102,
-0.12920638918876648,
-0.03334224596619606,
0.05622926726937294,
-0.036913324147462845,
0.12237885594367981,
-0.0848400890827179,
0.06590236723423004,
0.0033717823680490255,
-0.030056459829211235,
-0.0488768145442009,
0.07410939037799835,
0.0037821508012712,
0.0033065560273826122,
0.05448482558131218,
0.08353763818740845,
-0.10252245515584946,
0.22748969495296478,
0.022008858621120453,
-0.0552060529589653,
-0.012208325788378716,
0.023677172139286995,
0.07569369673728943,
0.08630989491939545,
0.0768752321600914,
0.04544786736369133,
0.0076964241452515125,
-0.03187931701540947,
-0.005878146272152662,
-0.12732483446598053,
-0.031544029712677,
0.0262546818703413,
-0.08584392815828323,
0.08573831617832184,
0.053216565400362015,
-0.07869172096252441,
0.037904106080532074,
-0.10425495356321335,
0.047830596566200256,
-0.00037045046337880194,
-0.03416381776332855,
-0.07438573241233826,
0.10739056020975113,
-0.11135460436344147,
-0.14164873957633972,
-0.1584516018629074,
0.07071401923894882,
-0.013750647194683552,
0.031653840094804764,
0.059720441699028015,
-0.059274956583976746,
-0.07650016248226166,
-0.12598484754562378,
0.00768353883177042,
0.015426871366798878,
-0.001579002826474607,
0.023609159514307976,
0.02989429607987404,
0.03271171823143959,
-0.09866704791784286,
0.0011142116272822022,
-0.019764743745326996,
-0.03773292899131775,
-0.013042556121945381,
-0.04109010100364685,
0.11747066676616669,
0.07815013080835342,
0.001458419137634337,
0.005885410122573376,
-0.010981082916259766,
0.12890011072158813,
-0.021640237420797348,
0.07742399722337723,
0.2640245854854584,
0.09644011408090591,
0.0708361342549324,
0.09710776805877686,
0.02183249406516552,
-0.05817316845059395,
0.07324230670928955,
-0.013621251098811626,
-0.1459822803735733,
-0.12938113510608673,
-0.12329518049955368,
-0.062488485127687454,
-0.03692504018545151,
0.050780124962329865,
0.0391671359539032,
0.07730250060558319,
0.12977290153503418,
-0.039218720048666,
0.02265484817326069,
0.05956631898880005,
0.08737419545650482,
-0.029885046184062958,
-0.021854707971215248,
0.04061104729771614,
-0.10554519295692444,
0.005703666713088751,
0.08824149519205093,
-0.01924901455640793,
0.18379299342632294,
-0.1196516901254654,
0.026231689378619194,
-0.002117378171533346,
0.015643596649169922,
0.0916099026799202,
0.07155746966600418,
-0.030447853729128838,
-0.005564871244132519,
-0.034544918686151505,
-0.12391458451747894,
0.022090338170528412,
0.13509337604045868,
-0.037337251007556915,
0.07670941203832626,
-0.01341547816991806,
0.06484012305736542,
0.014956587925553322,
0.013668136671185493,
0.03425007313489914,
-0.2635102868080139,
-0.02250114269554615,
0.0862514078617096,
0.11696454137563705,
-0.018797164782881737,
-0.009437420405447483,
0.25501546263694763,
-0.025511208921670914,
0.07597631961107254,
-0.024760186672210693,
0.0708763375878334,
0.031974535435438156,
-0.06626597046852112,
-0.01756204478442669,
0.18850736320018768,
-0.022618262097239494,
0.027616536244750023,
-0.1794394999742508,
0.006807814352214336,
0.024860532954335213,
0.10763569921255112,
-0.050709083676338196,
0.016758982092142105,
0.07770445942878723,
0.08711922913789749,
0.10357018560171127,
-0.001675178064033389,
-0.0433010533452034,
-0.13892044126987457,
-0.09195899963378906,
-0.007932925596833229,
0.0268175657838583,
-0.04780730605125427,
0.040290407836437225,
0.007990965619683266,
-0.025260690599679947,
0.0010236927773803473,
-0.042241837829351425,
-0.10830533504486084,
-0.13176389038562775,
-0.026668572798371315,
0.2141551524400711,
0.014840657822787762,
-0.08865337073802948,
-0.06456120312213898,
0.02750895731151104,
0.055813707411289215,
-0.03200140222907066,
-0.10954766720533371,
-0.1039484441280365,
-0.06225476786494255,
0.1044953465461731,
-0.05319923162460327,
0.02457553520798683,
0.004609039984643459,
0.12013374269008636,
-0.10124830901622772,
-0.09990958869457245,
-0.016005488112568855,
-0.09544211626052856,
-0.12083582580089569,
-0.05573583021759987,
0.0622376948595047,
-0.0038727549836039543,
0.018684426322579384,
-0.0043965294025838375,
0.014362887479364872,
0.05405285954475403,
-0.1018320843577385,
-0.023456856608390808,
0.2867535650730133,
-0.030277660116553307,
0.17485028505325317,
-0.07638189941644669,
-0.1216762587428093,
-0.023332808166742325,
-0.0040689753368496895,
0.018446579575538635,
0.18534210324287415,
-0.09551016241312027,
0.06467088311910629,
0.08797735720872879,
-0.08552055060863495,
-0.22420430183410645,
0.021649427711963654,
0.015957260504364967,
0.03384280949831009,
0.026998478919267654,
-0.14675110578536987,
0.13361364603042603,
0.03587627038359642,
-0.022679517045617104,
0.16402409970760345,
-0.313018798828125,
-0.11843345314264297,
-0.041912924498319626,
0.1702507585287094,
0.1573439985513687,
-0.13255158066749573,
-0.03270656615495682,
-0.06347974389791489,
0.0288421381264925,
0.10486374795436859,
-0.1373053789138794,
0.09403892606496811,
0.0027441882994025946,
-0.05961122363805771,
0.03180977329611778,
-0.05598610267043114,
0.15229466557502747,
-0.05588935315608978,
0.08231724053621292,
-0.04012461379170418,
0.06167663633823395,
0.09183156490325928,
-0.08033566921949387,
0.10971338301897049,
-0.2025560885667801,
0.048021066933870316,
-0.04540081322193146,
-0.03453189507126808,
0.024300819262862206,
0.05212109163403511,
-0.0026281478349119425,
-0.0719948559999466,
-0.05014609917998314,
0.06121525540947914,
0.07718884199857712,
-0.00003423326779739,
-0.03927423432469368,
-0.0428522489964962,
0.04178551211953163,
0.17405974864959717,
0.0028169704601168633,
0.036997538059949875,
0.021275632083415985,
0.013039137236773968,
-0.044696252793073654,
0.11519553512334824,
-0.1474103480577469,
0.004402145743370056,
0.07603348791599274,
0.04489549621939659,
0.1332227736711502,
0.015222553163766861,
-0.0690051019191742,
0.07447570562362671,
0.07281693816184998,
-0.07258850336074829,
-0.1248074620962143,
-0.0622234009206295,
-0.011682173237204552,
-0.02293456718325615,
-0.016501912847161293,
0.12494106590747833,
-0.10476090013980865,
0.03728315606713295,
-0.026456156745553017,
0.04545627906918526,
-0.023228222504258156,
0.12179102748632431,
0.022318968549370766,
-0.008508766070008278,
-0.06774233281612396,
0.08442836254835129,
0.03133062273263931,
-0.010108007118105888,
0.0343901589512825,
0.08196725696325302,
-0.07653649151325226,
-0.042940594255924225,
-0.019660532474517822,
0.1364019513130188,
-0.01427333615720272,
-0.0579545758664608,
-0.10992537438869476,
-0.08795510977506638,
-0.05041373148560524,
0.131325364112854,
0.07156982272863388,
-0.02771761454641819,
-0.03646356612443924,
-0.04462538659572601,
-0.10368119180202484,
0.09875413030385971,
0.0450555682182312,
0.08136405795812607,
-0.1945866346359253,
0.013678362593054771,
0.05028225854039192,
-0.052338432520627975,
-0.05886056274175644,
-0.025689151138067245,
-0.0751139372587204,
-0.005206389352679253,
-0.04426164552569389,
0.1458960473537445,
-0.06729815155267715,
0.019158203154802322,
-0.030321821570396423,
-0.02192031219601631,
-0.00006743729318259284,
0.07088939845561981,
-0.012163168750703335,
-0.015147690661251545,
-0.014954330399632454,
0.04470520466566086,
-0.07424546778202057,
-0.10266188532114029,
-0.029990937560796738,
-0.06462064385414124,
-0.012884477153420448,
-0.021070249378681183,
-0.061122823506593704,
0.013143200427293777,
-0.23758448660373688,
0.04197045415639877,
0.1228213682770729,
0.02901914343237877,
-0.016147788614034653,
-0.023768732324242592,
0.01743593066930771,
-0.005144428927451372,
0.005690992809832096,
-0.02564830705523491,
0.0011657411232590675,
-0.08697427064180374,
0.009582075290381908,
-0.05359897017478943,
0.009033594280481339,
-0.050830841064453125,
0.057002000510692596,
0.1747191995382309,
0.10198234021663666,
0.1142701581120491,
-0.1039578914642334,
0.08706224709749222,
-0.1117260679602623,
-0.022594336420297623,
-0.018108736723661423,
-0.06725170463323593,
-0.09659796953201294,
-0.024313347414135933,
0.0213736891746521,
-0.00348881003446877,
0.0841403380036354,
0.04151752218604088,
-0.030575444921851158,
-0.03020479343831539,
-0.037926919758319855,
0.11389324814081192,
-0.018390493467450142,
0.19207344949245453,
0.017928805202245712,
0.023083217442035675,
-0.0327627956867218,
0.04183030501008034,
0.10339544713497162,
0.13161952793598175,
0.007832396775484085,
0.11985710263252258,
0.0061972299590706825,
0.08482375741004944,
-0.011663445271551609,
0.023689888417720795,
-0.034209344536066055,
0.06329595297574997,
-0.0537073090672493,
0.03180764243006706,
-0.05005961284041405,
0.021594634279608727,
0.0854622945189476,
-0.1261260211467743,
-0.015580597333610058,
0.11512570828199387,
-0.047284677624702454,
-0.12069080024957657,
-0.1683165729045868,
-0.08406107872724533,
-0.14033177495002747,
-0.00877372920513153,
-0.07939771562814713,
0.01907620020210743,
0.038954831659793854,
-0.028709886595606804,
0.04765394702553749,
0.16588647663593292,
-0.03318066895008087,
-0.0007371503161266446,
0.05808515474200249,
-0.030080797150731087,
-0.020561855286359787,
0.04766552522778511,
-0.04181372746825218,
0.11685311049222946,
-0.031285811215639114,
-0.004135673865675926,
0.03810425102710724,
0.03521329537034035,
0.04823397472500801,
-0.010794143192470074,
-0.0782947838306427,
-0.018317829817533493,
0.034842055290937424,
-0.0055892737582325935,
0.19660019874572754,
0.1067984402179718,
-0.05527224764227867,
-0.019654853269457817,
0.14522549510002136,
-0.06212960556149483,
-0.029445722699165344,
-0.12235436588525772,
0.14222915470600128,
-0.04464511573314667,
0.011741225607693195,
-0.004849080927670002,
-0.09322208911180496,
-0.017739204689860344,
0.10115375369787216,
0.14256347715854645,
0.019839761778712273,
0.03173590824007988,
-0.10150495171546936,
-0.019816678017377853,
-0.07589669525623322,
0.06260097026824951,
0.05455348640680313,
0.19338358938694,
-0.016012465581297874,
0.05403687804937363,
-0.046658966690301895,
-0.05665302276611328,
-0.0951433852314949,
-0.08846032619476318,
-0.08452765643596649,
-0.027811171486973763,
-0.016619505360722542,
0.035841792821884155,
-0.06767157465219498,
-0.07853778451681137,
0.18596699833869934,
-0.10732121020555496,
-0.029335716739296913,
-0.06782299280166626,
0.029149362817406654,
0.04615508019924164,
-0.02201167494058609,
-0.08869714289903641,
0.02018994651734829,
0.048301976174116135,
-0.03557905554771423,
-0.10374891757965088,
-0.08082457631826401,
-0.04189296439290047,
-0.14067958295345306,
0.16182449460029602,
-0.03200952708721161,
-0.013250673189759254,
0.04010753706097603,
0.0002631675451993942,
-0.10364048182964325,
0.12113496661186218,
-0.029223749414086342,
-0.11247824132442474,
-0.008313897997140884,
0.16558019816875458,
-0.11149181425571442,
0.2096109688282013,
0.022971734404563904,
-0.1056864783167839,
0.01891201362013817,
0.040444355458021164,
-0.0708513930439949,
-0.08087588846683502,
-0.0162029005587101,
-0.12242056429386139,
0.12639567255973816,
0.12923648953437805,
-0.009103535674512386,
-0.01941562071442604,
0.002939272206276655,
0.05208928510546684,
0.03829574212431908,
0.013665902428328991,
0.05763723701238632,
-0.08780422061681747,
-0.0521051324903965,
0.02552456595003605,
0.05823734030127525,
-0.2541719973087311,
-0.03465826436877251,
-0.17217007279396057,
-0.051500607281923294,
-0.012263593263924122,
0.03131452947854996,
0.17814505100250244,
0.023856284096837044,
-0.027030421420931816,
-0.2931806147098541,
0.026213308796286583,
0.11417213082313538,
-0.08187741786241531,
-0.07213529944419861
] |
null | null | transformers | # miqu-1-120b

* EXL2: [2.4bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2) | [2.65bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2) | 3.0bpw | [4.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2) | [5.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2)
* GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/) | [IQ3_XXS](https://huggingface.co/wolfram/miqu-1-120b-GGUF)
* HF FP16: [wolfram/miqu-1-120b](https://huggingface.co/wolfram/miqu-1-120b)
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
Inspired by [Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2), [MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [goliath-120b](https://huggingface.co/alpindale/goliath-120b).
Thanks for the support, [CopilotKit](https://github.com/CopilotKit/CopilotKit) - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, [Lone Striker](https://huggingface.co/LoneStriker)!
## Prompt template: Mistral
```
<s>[INST] {prompt} [/INST]
```
See also: [🐺🐦⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates : LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/18ljvxb/llm_prompt_format_comparisontest_mixtral_8x7b/)
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
dtype: float16
merge_method: passthrough
slices:
- sources:
- layer_range: [0, 20]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [10, 30]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [20, 40]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [30, 50]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [40, 60]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [50, 70]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [60, 80]
model: 152334H/miqu-1-70b-sf
```
## Credits & Special Thanks
* original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
* leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
* f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
* mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
* mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
### Support
* [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| {"language": ["en", "de", "fr", "es", "it"], "library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["152334H/miqu-1-70b-sf"]} | text-generation | LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"en",
"de",
"fr",
"es",
"it",
"base_model:152334H/miqu-1-70b-sf",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:27:12+00:00 | [] | [
"en",
"de",
"fr",
"es",
"it"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # miqu-1-120b
!image/jpeg
* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw
* GGUF: Q2_K-Q5_K_M | IQ3_XXS
* HF FP16: wolfram/miqu-1-120b
This is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.
Inspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.
Thanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, Lone Striker!
## Prompt template: Mistral
See also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* 152334H/miqu-1-70b-sf
### Configuration
The following YAML configuration was used to produce this model:
## Credits & Special Thanks
* original (unreleased) model: mistralai (Mistral AI_)
* leaked model: miqudev/miqu-1-70b
* f16 model: 152334H/miqu-1-70b-sf
* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.
* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2
### Support
* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| [
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
85,
206,
44,
31,
4,
17,
28,
17,
107,
69,
43
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140## Merge Details### Merge Method\n\nThis model was merged using the passthrough merge method.### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.08452296257019043,
0.07996358722448349,
-0.006216804962605238,
0.07489742338657379,
0.062447689473629,
0.048032015562057495,
0.10850238054990768,
0.11730548739433289,
0.04997159540653229,
0.1304655373096466,
0.019059622660279274,
0.0882313996553421,
0.10314993560314178,
0.1546645164489746,
-0.0008784975507296622,
-0.13700617849826813,
0.03406155854463577,
-0.032812073826789856,
-0.05265757441520691,
0.07489487528800964,
0.07002948969602585,
-0.049538981169462204,
0.07827795296907425,
0.01992727816104889,
-0.0794241651892662,
-0.02351442351937294,
-0.03350690379738808,
-0.0034194772597402334,
0.06382977962493896,
0.09272785484790802,
0.010276054963469505,
0.01770416460931301,
0.010751184076070786,
-0.1683615893125534,
0.032938648015260696,
0.03592948243021965,
-0.017932593822479248,
0.07718798518180847,
0.10788709670305252,
-0.018908003345131874,
0.021088771522045135,
-0.10027661174535751,
-0.02568136155605316,
0.08249933272600174,
-0.10719147324562073,
-0.1474289745092392,
-0.18762712180614471,
0.15491966903209686,
0.042507048696279526,
0.03724713996052742,
-0.011902958154678345,
0.1333865076303482,
0.050249699503183365,
0.06016452983021736,
0.24593202769756317,
-0.22131793200969696,
-0.061919860541820526,
0.028957664966583252,
0.05541115999221802,
-0.04985284432768822,
-0.021356018260121346,
0.01992829702794552,
0.04054830223321915,
0.02191905304789543,
-0.020568734034895897,
-0.04234932363033295,
0.13632027804851532,
-0.05843184515833855,
-0.13171182572841644,
-0.02820662222802639,
0.09477535635232925,
0.04355064406991005,
-0.060202740132808685,
-0.1276080310344696,
-0.0682656466960907,
-0.03270931541919708,
-0.0022935839369893074,
-0.04131745919585228,
-0.002290143631398678,
-0.011816466227173805,
0.08523242175579071,
-0.08333887904882431,
-0.05337991565465927,
-0.020235886797308922,
-0.06682160496711731,
0.16135647892951965,
0.018354052677750587,
0.013679140247404575,
-0.013255320489406586,
0.07744071632623672,
-0.1265559196472168,
-0.1336318850517273,
-0.05428304150700569,
-0.03462158143520355,
-0.05692019686102867,
-0.05218517407774925,
0.01661340333521366,
-0.07750479131937027,
0.08107020705938339,
0.175857275724411,
-0.11741548776626587,
0.06220667064189911,
-0.009697522968053818,
0.030903007835149765,
0.024629943072795868,
0.08424429595470428,
-0.08733733743429184,
-0.08122188597917557,
0.04335138201713562,
0.05078619346022606,
0.05994807183742523,
0.0019238290842622519,
0.00010135882621398196,
-0.02575274370610714,
0.0019080113852396607,
0.07156599313020706,
0.06952427327632904,
0.029285242781043053,
-0.07679720222949982,
-0.05144278332591057,
0.17693321406841278,
-0.10877041518688202,
0.05875063315033913,
0.038095518946647644,
-0.010931058786809444,
-0.015534471720457077,
0.026578398421406746,
-0.028613878414034843,
-0.05422482639551163,
0.014574683271348476,
-0.052837468683719635,
0.006087095942348242,
-0.04121243208646774,
-0.04512706398963928,
0.052960656583309174,
0.044456131756305695,
-0.05848181992769241,
-0.12249575555324554,
-0.08548451960086823,
-0.056048691272735596,
0.033982738852500916,
-0.07165446132421494,
0.009947382844984531,
0.03666377440094948,
-0.08962082862854004,
0.029198521748185158,
0.02603435143828392,
0.022042686119675636,
-0.022196292877197266,
0.013299040496349335,
0.01655365526676178,
0.02530047670006752,
-0.031476203352212906,
-0.009267413057386875,
-0.05801313742995262,
0.08978775143623352,
-0.19187140464782715,
0.07319818437099457,
-0.06753759831190109,
-0.0458916574716568,
-0.1367833912372589,
-0.022929007187485695,
0.0001557070208946243,
-0.021685365587472916,
0.08853033930063248,
0.14028236269950867,
-0.15254990756511688,
-0.01800709031522274,
0.133183553814888,
-0.09939759224653244,
-0.08947071433067322,
0.09990517795085907,
0.010765177197754383,
-0.005259351339191198,
0.010271748527884483,
0.08585681021213531,
0.12536770105361938,
-0.056073881685733795,
-0.0745566338300705,
-0.10543492436408997,
-0.02353603020310402,
0.10377749055624008,
0.036487940698862076,
-0.06648851186037064,
0.006608173716813326,
0.0013413212727755308,
-0.03926616534590721,
-0.02470051497220993,
-0.01713505946099758,
-0.04516416788101196,
-0.051653604954481125,
-0.03216031193733215,
0.019755378365516663,
-0.0311356820166111,
-0.04306983947753906,
-0.06874691694974899,
-0.0662546455860138,
-0.03937284275889397,
0.10621660202741623,
0.0002958594122901559,
0.013465981930494308,
-0.06993132084608078,
0.12132570147514343,
-0.043282829225063324,
0.019509179517626762,
-0.13229045271873474,
-0.062326252460479736,
0.05427157133817673,
-0.17642687261104584,
0.057915039360523224,
-0.07802051305770874,
0.07007356733083725,
0.08792263269424438,
-0.03142979368567467,
-0.04443249851465225,
0.014312747865915298,
-0.011503690853714943,
-0.05503445118665695,
-0.13744132220745087,
-0.05339666083455086,
-0.03718583658337593,
0.1360798180103302,
-0.07717297226190567,
0.03823299705982208,
0.04849797114729881,
0.17298123240470886,
0.013948954641819,
-0.06230805069208145,
0.059276606887578964,
0.013416923582553864,
-0.01901276968419552,
-0.04254152253270149,
0.022224929183721542,
-0.015195275656878948,
-0.06335578858852386,
0.06797880679368973,
-0.16682805120944977,
-0.09226492792367935,
0.08205856382846832,
0.11543388664722443,
-0.10825490206480026,
-0.0011853284668177366,
-0.0016989412251859903,
-0.03144387900829315,
-0.02349192649126053,
-0.07283974438905716,
0.139337420463562,
0.03313908725976944,
0.05388158559799194,
-0.05538925528526306,
-0.07884406298398972,
0.012225146405398846,
-0.024691954255104065,
-0.07743193954229355,
0.12690365314483643,
0.02291349321603775,
-0.19036413729190826,
0.1086549460887909,
0.10671419650316238,
0.024346476420760155,
0.12972186505794525,
0.010163333266973495,
-0.020192109048366547,
-0.11323220282793045,
0.07214321941137314,
0.06328365206718445,
0.08764441311359406,
0.009260048158466816,
0.06931837648153305,
0.03807743266224861,
0.00207220995798707,
0.013925828970968723,
-0.11039725691080093,
0.03326145187020302,
0.019972439855337143,
-0.035299137234687805,
0.07873526960611343,
0.04182936251163483,
0.027880189940333366,
0.08705969154834747,
0.017179058864712715,
0.07423700392246246,
-0.01186748594045639,
-0.03507009521126747,
-0.09978814423084259,
0.13680647313594818,
-0.10421305149793625,
-0.1235826313495636,
-0.15689323842525482,
0.023495908826589584,
-0.1045532375574112,
-0.031237754970788956,
0.04314960539340973,
-0.04964744672179222,
-0.04190392047166824,
-0.09863945841789246,
0.07575149089097977,
0.04664899781346321,
-0.031183557584881783,
0.029061678797006607,
-0.04268911853432655,
0.05956493318080902,
-0.1079893708229065,
-0.03650837019085884,
0.0311584435403347,
0.003986986353993416,
0.03826166316866875,
0.06534901261329651,
0.09445244818925858,
0.06816873699426651,
0.010970757342875004,
-0.007111722603440285,
0.0249604694545269,
0.2451404184103012,
-0.09102492034435272,
0.08363116532564163,
0.1731153130531311,
0.012367261573672295,
0.08963974565267563,
0.1418784111738205,
0.0519220232963562,
-0.06685937941074371,
-0.013835576362907887,
0.023638667538762093,
-0.024145755916833878,
-0.19078323245048523,
-0.061981018632650375,
-0.049455784261226654,
0.02179974503815174,
0.016594912856817245,
0.056652046740055084,
-0.03246590495109558,
0.04957956075668335,
-0.06632956862449646,
-0.023145724087953568,
0.04272812604904175,
0.07885605841875076,
0.04942423850297928,
0.05232582241296768,
0.032079048454761505,
-0.030630147084593773,
-0.01874661073088646,
0.10800144821405411,
0.002024435205385089,
0.0828050896525383,
0.005763496737927198,
0.15540826320648193,
0.042188387364149094,
0.05763785168528557,
-0.006694222800433636,
0.05031910166144371,
0.012587900273501873,
0.01727849431335926,
0.011400274932384491,
-0.0922514945268631,
0.017045721411705017,
0.037224024534225464,
0.031022431328892708,
0.08605832606554031,
-0.019477766007184982,
-0.042653512209653854,
0.06828916072845459,
0.20636118948459625,
0.054764457046985626,
-0.18104644119739532,
-0.052608687430620193,
0.05323893949389458,
-0.005172972101718187,
-0.0441189780831337,
-0.0574423223733902,
0.0601525716483593,
-0.1421785205602646,
0.10297753661870956,
-0.02837306633591652,
0.08344317972660065,
-0.0768112912774086,
-0.025202374905347824,
-0.002551008015871048,
0.14314106106758118,
-0.007685830816626549,
0.009870217181742191,
-0.11473074555397034,
0.09677925705909729,
0.047673650085926056,
0.050363969057798386,
-0.0467650331556797,
0.05014784261584282,
0.06751372665166855,
-0.016121957451105118,
0.1009698212146759,
0.015580941922962666,
-0.08597638458013535,
-0.13121242821216583,
-0.07546716183423996,
-0.02328302152454853,
0.09724711626768112,
-0.07956860214471817,
0.11797893047332764,
-0.006677214056253433,
-0.041914381086826324,
-0.051285892724990845,
0.08619783818721771,
-0.16726025938987732,
-0.10657192021608353,
0.08049735426902771,
0.035933516919612885,
-0.00293326354585588,
-0.07889305055141449,
-0.03560107201337814,
-0.06792481988668442,
0.20780132710933685,
-0.12493043392896652,
-0.06458897888660431,
-0.12241725623607635,
-0.02887725457549095,
0.14838698506355286,
-0.09519194811582565,
0.016095273196697235,
-0.0620892271399498,
0.09675031155347824,
-0.05750473216176033,
-0.09322291612625122,
0.03633593022823334,
-0.0772782415151596,
-0.22163733839988708,
-0.029905345290899277,
0.1838245987892151,
0.031322285532951355,
0.03489852324128151,
0.0206821970641613,
0.04595774784684181,
0.021382097154855728,
-0.10996323078870773,
0.04105592891573906,
0.12817256152629852,
0.020900849252939224,
0.09039270132780075,
-0.045222893357276917,
-0.14004118740558624,
-0.09013182669878006,
-0.05677556246519089,
0.0997968390583992,
0.29861730337142944,
-0.05223260074853897,
0.030688846483826637,
0.04147891700267792,
-0.07043230533599854,
-0.15785279870033264,
-0.038915667682886124,
0.04657349735498428,
0.01753225363790989,
0.01858132891356945,
-0.07110119611024857,
0.0366257019340992,
0.10373353213071823,
-0.01873767003417015,
0.15667074918746948,
-0.263369619846344,
-0.12704867124557495,
-0.006682985462248325,
0.06700930744409561,
-0.01957916095852852,
-0.1452975571155548,
-0.11674544960260391,
-0.03796976059675217,
-0.1995268017053604,
0.09981389343738556,
-0.02467156946659088,
0.0965239554643631,
-0.023898916319012642,
0.015778150409460068,
0.03433018922805786,
-0.059150923043489456,
0.19815395772457123,
-0.039659541100263596,
0.04237253591418266,
-0.08535932749509811,
0.011777906678617,
0.026407601311802864,
-0.05703800171613693,
0.0950261577963829,
-0.035693515092134476,
0.017066331580281258,
-0.07961391657590866,
0.0030619590543210506,
-0.08434901386499405,
0.06988205015659332,
-0.053291499614715576,
-0.002292127348482609,
-0.06203543022274971,
0.07322024554014206,
0.02764309197664261,
-0.02415998838841915,
0.024202430620789528,
-0.021172653883695602,
0.10087262094020844,
0.2350863218307495,
0.09775305539369583,
0.05048970505595207,
-0.06141071021556854,
-0.00026774543221108615,
-0.04018312692642212,
0.02579716220498085,
-0.035933900624513626,
0.04462524875998497,
0.08787034451961517,
0.02455063909292221,
0.10782863199710846,
0.01810472458600998,
-0.1601009964942932,
0.018065448850393295,
0.06646080315113068,
-0.13265873491764069,
-0.2303391844034195,
0.003492377931252122,
0.04154038056731224,
-0.05837193876504898,
0.01806606538593769,
0.1862204670906067,
0.0009367137681692839,
-0.04963487386703491,
0.022825637832283974,
0.05416732281446457,
-0.03536570817232132,
0.15667866170406342,
0.00837104581296444,
0.04538920149207115,
-0.061773963272571564,
0.07580430060625076,
0.04042122885584831,
0.0016077554319053888,
0.018714869394898415,
0.17067234218120575,
-0.05190492421388626,
-0.07857772707939148,
0.04102415591478348,
0.03600002080202103,
-0.021241383627057076,
-0.021828800439834595,
-0.027304736897349358,
-0.10624261945486069,
0.013062734156847,
0.12654060125350952,
0.02681851200759411,
-0.018991822376847267,
0.08323623985052109,
-0.004140008706599474,
-0.028763674199581146,
0.07244459539651871,
0.05491362512111664,
0.08497325330972672,
-0.059819940477609634,
0.06284529715776443,
-0.02697847969830036,
0.03915136307477951,
-0.0026654857210814953,
0.019256316125392914,
-0.12592482566833496,
-0.05442740023136139,
-0.08961838483810425,
0.005896841175854206,
-0.11454344540834427,
-0.029031215235590935,
-0.0207052081823349,
-0.00936585757881403,
-0.02115669846534729,
0.03387326002120972,
-0.051367443054914474,
-0.11562348157167435,
-0.07470326870679855,
0.08342844247817993,
-0.14418748021125793,
-0.01641484536230564,
0.06919168680906296,
-0.07516539096832275,
0.09250383824110031,
0.01708120107650757,
0.0034909669775515795,
-0.009268661960959435,
-0.11126343160867691,
-0.02617836929857731,
-0.013749503530561924,
0.020847713574767113,
0.02727784588932991,
-0.20117603242397308,
-0.002813884522765875,
-0.04270019009709358,
-0.03368119150400162,
0.01241485308855772,
0.15425018966197968,
-0.09736622869968414,
-0.040856342762708664,
-0.03006616234779358,
-0.044825147837400436,
-0.06249745562672615,
0.018799860030412674,
0.10122360289096832,
0.009257921017706394,
0.08748460561037064,
-0.05421772226691246,
0.07078760117292404,
-0.17954121530056,
-0.01305952575057745,
0.022268623113632202,
-0.06352261453866959,
0.042893052101135254,
0.006506424397230148,
0.07432542741298676,
-0.024414660409092903,
0.033989157527685165,
-0.08212705701589584,
-0.018906600773334503,
0.026504697278141975,
-0.028520310297608376,
-0.021283207461237907,
0.017214683815836906,
0.020455483347177505,
-0.022489050403237343,
-0.024124853312969208,
-0.0313778817653656,
0.020932979881763458,
-0.019497908651828766,
-0.05568772181868553,
0.1307651400566101,
0.10116901993751526,
0.042394958436489105,
0.06033344939351082,
0.0699099600315094,
-0.06935380399227142,
-0.019917812198400497,
-0.015662092715501785,
-0.0501786470413208,
0.05921950191259384,
-0.0570748895406723,
0.11598431318998337,
0.0971740260720253,
-0.16128812730312347,
0.08817395567893982,
-0.03033575415611267,
-0.025268414989113808,
-0.0654015764594078,
-0.1747601330280304,
-0.06647072732448578,
-0.021373050287365913,
-0.007110625971108675,
-0.07304833084344864,
0.04277768358588219,
0.03683268278837204,
0.0017412106972187757,
-0.0061371480114758015,
0.12685923278331757,
-0.04393889009952545,
-0.058910902589559555,
0.035390980541706085,
0.037289973348379135,
-0.015837162733078003,
0.04472239688038826,
-0.006923268549144268,
0.022451741620898247,
0.012978330254554749,
0.05661128833889961,
0.08223818242549896,
-0.00805328506976366,
0.02584826387465,
-0.028271563351154327,
-0.10848145186901093,
-0.0006025367183610797,
0.005287060514092445,
0.01833254098892212,
0.08337931334972382,
0.05019523203372955,
0.016249405220150948,
-0.015350534580647945,
0.0810377299785614,
-0.02552863396704197,
-0.014691580086946487,
-0.1009829118847847,
0.15841826796531677,
-0.0559045635163784,
0.002254345454275608,
-0.0015080315060913563,
-0.11343378573656082,
-0.011597293429076672,
0.12920117378234863,
0.16337698698043823,
-0.08000685274600983,
0.03376281261444092,
0.017094949260354042,
0.00689410325139761,
0.004397694021463394,
0.049244899302721024,
0.059573933482170105,
0.10981784015893936,
-0.046494387090206146,
0.12055498361587524,
-0.0517595037817955,
-0.032632842659950256,
-0.07947799563407898,
0.09397511184215546,
-0.047008223831653595,
0.014640634879469872,
-0.017156826332211494,
0.07035721093416214,
-0.032913386821746826,
-0.16172361373901367,
0.03033449873328209,
-0.12312323600053787,
-0.138545423746109,
-0.00011208858632016927,
-0.0077992831356823444,
0.001197756384499371,
0.07809285819530487,
0.009989638812839985,
0.010767842642962933,
0.13262617588043213,
0.004294542595744133,
-0.07419104129076004,
-0.056784018874168396,
0.0005215644487179816,
-0.09722831845283508,
0.18984216451644897,
0.024254996329545975,
0.04178498685359955,
0.13091936707496643,
-0.033293675631284714,
-0.15407414734363556,
0.05585271492600441,
0.0700296014547348,
-0.09110260009765625,
0.022162333130836487,
0.14707012474536896,
-0.005258307326585054,
0.030615337193012238,
0.07387614995241165,
-0.053392838686704636,
0.019069518893957138,
0.0722138062119484,
0.019377747550606728,
-0.10239636152982712,
0.08977871388196945,
-0.10703950375318527,
0.13002605736255646,
0.18838851153850555,
-0.03489691764116287,
0.012865101918578148,
-0.03789073973894119,
0.04828639328479767,
0.038723692297935486,
0.09065895527601242,
-0.034608691930770874,
-0.19317959249019623,
0.06994378566741943,
-0.023887865245342255,
0.08475272357463837,
-0.1766246110200882,
-0.1239243745803833,
-0.029406068846583366,
-0.021863967180252075,
-0.013726417906582355,
0.0984341949224472,
0.09909792244434357,
0.0073255919851362705,
-0.023081263527274132,
-0.1434156447649002,
-0.02458539977669716,
0.10146303474903107,
-0.08631070703268051,
-0.04105154797434807
] |
null | null | transformers |
# MarkrAI/RAG-KO-Mixtral-7Bx2-v2.0
# Model Details
## Model Developers
MarkrAI - AI Researchers
## Base Model
[DopeorNope/Ko-Mixtral-v1.4-MoE-7Bx2](https://huggingface.co/DopeorNope/Ko-Mixtral-v1.4-MoE-7Bx2).
## Instruction tuning Method
Using QLoRA.
```
4-bit quantization
Lora_r: 64
Lora_alpha: 64
Lora_dropout: 0.05
Lora_target_modules: [embed_tokens, q_proj, k_proj, v_proj, o_proj, gate, w1, w2, w3, lm_head]
```
## Hyperparameters
```
Epoch: 5
Batch size: 64
Learning_rate: 1e-5
Learning scheduler: linear
Warmup_ratio: 0.06
```
## Datasets
Private datasets: [HumanF-MarkrAI/Korean-RAG-ver2](https://huggingface.co/datasets/HumanF-MarkrAI/Korean-RAG-ver2)
```
Aihub datasets 활용하여서 제작함.
```
## Implmentation Code
```
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
repo = "MarkrAI/RAG-KO-Mixtral-7Bx2-v2.0"
markrAI_RAG = AutoModelForCausalLM.from_pretrained(
repo,
return_dict=True,
torch_dtype=torch.float16,
device_map='auto'
)
markrAI_RAG_tokenizer = AutoTokenizer.from_pretrained(repo)
```
# Model Benchmark
- Coming soon... | {"language": ["ko"], "license": "cc-by-nc-sa-4.0", "tags": ["Retrieval Augmented Generation", "RAG", "Multi-domain"], "datasets": ["HumanF-MarkrAI/Korean-RAG-ver2"]} | text-generation | MarkrAI/RAG-KO-Mixtral-7Bx2-v2.0 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"Retrieval Augmented Generation",
"RAG",
"Multi-domain",
"ko",
"dataset:HumanF-MarkrAI/Korean-RAG-ver2",
"license:cc-by-nc-sa-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:28:22+00:00 | [] | [
"ko"
] | TAGS
#transformers #safetensors #mixtral #text-generation #Retrieval Augmented Generation #RAG #Multi-domain #ko #dataset-HumanF-MarkrAI/Korean-RAG-ver2 #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# MarkrAI/RAG-KO-Mixtral-7Bx2-v2.0
# Model Details
## Model Developers
MarkrAI - AI Researchers
## Base Model
DopeorNope/Ko-Mixtral-v1.4-MoE-7Bx2.
## Instruction tuning Method
Using QLoRA.
## Hyperparameters
## Datasets
Private datasets: HumanF-MarkrAI/Korean-RAG-ver2
## Implmentation Code
# Model Benchmark
- Coming soon... | [
"# MarkrAI/RAG-KO-Mixtral-7Bx2-v2.0",
"# Model Details",
"## Model Developers \nMarkrAI - AI Researchers",
"## Base Model \nDopeorNope/Ko-Mixtral-v1.4-MoE-7Bx2.",
"## Instruction tuning Method \nUsing QLoRA.",
"## Hyperparameters",
"## Datasets\nPrivate datasets: HumanF-MarkrAI/Korean-RAG-ver2",
"## Implmentation Code",
"# Model Benchmark\n- Coming soon..."
] | [
"TAGS\n#transformers #safetensors #mixtral #text-generation #Retrieval Augmented Generation #RAG #Multi-domain #ko #dataset-HumanF-MarkrAI/Korean-RAG-ver2 #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MarkrAI/RAG-KO-Mixtral-7Bx2-v2.0",
"# Model Details",
"## Model Developers \nMarkrAI - AI Researchers",
"## Base Model \nDopeorNope/Ko-Mixtral-v1.4-MoE-7Bx2.",
"## Instruction tuning Method \nUsing QLoRA.",
"## Hyperparameters",
"## Datasets\nPrivate datasets: HumanF-MarkrAI/Korean-RAG-ver2",
"## Implmentation Code",
"# Model Benchmark\n- Coming soon..."
] | [
98,
19,
3,
11,
24,
12,
5,
24,
6,
10
] | [
"passage: TAGS\n#transformers #safetensors #mixtral #text-generation #Retrieval Augmented Generation #RAG #Multi-domain #ko #dataset-HumanF-MarkrAI/Korean-RAG-ver2 #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MarkrAI/RAG-KO-Mixtral-7Bx2-v2.0# Model Details## Model Developers \nMarkrAI - AI Researchers## Base Model \nDopeorNope/Ko-Mixtral-v1.4-MoE-7Bx2.## Instruction tuning Method \nUsing QLoRA.## Hyperparameters## Datasets\nPrivate datasets: HumanF-MarkrAI/Korean-RAG-ver2## Implmentation Code# Model Benchmark\n- Coming soon..."
] | [
-0.03690986707806587,
0.13318860530853271,
-0.00145150744356215,
-0.03523612394928932,
0.1257423609495163,
-0.029501287266612053,
0.08868687599897385,
0.08752609044313431,
-0.028227047994732857,
0.07319030910730362,
0.1198701411485672,
0.058575715869665146,
0.054031625390052795,
0.19805492460727692,
-0.02174859680235386,
-0.33433428406715393,
0.0408770926296711,
-0.01524751354008913,
0.014891104772686958,
0.09017856419086456,
0.11343550682067871,
-0.03711149841547012,
0.12338407337665558,
0.021789709106087685,
-0.07697348296642303,
-0.05175556614995003,
-0.05195220187306404,
-0.09197578579187393,
0.04158615320920944,
0.03298555314540863,
0.05350165441632271,
0.06575577706098557,
0.036042533814907074,
-0.1252586543560028,
0.034875381737947464,
0.006527703255414963,
-0.004353903234004974,
0.09453416615724564,
0.038432564586400986,
-0.06830030679702759,
0.18973462283611298,
-0.09069894999265671,
-0.0129036083817482,
0.040473874658346176,
-0.07817400991916656,
-0.10817964375019073,
-0.12491483241319656,
0.023349633440375328,
0.10367731750011444,
0.04962414130568504,
-0.007453262340277433,
0.09623797237873077,
0.004205082543194294,
0.045561473816633224,
0.17242790758609772,
-0.34304091334342957,
-0.04980660602450371,
0.08022461086511612,
0.09650954604148865,
0.05952499434351921,
-0.06022517755627632,
0.06359317153692245,
0.050003502517938614,
0.04382293298840523,
0.07474418729543686,
-0.047989096492528915,
0.01864052750170231,
0.003422268433496356,
-0.12162969261407852,
0.026695983484387398,
0.24228882789611816,
0.03351780027151108,
-0.005275281611829996,
-0.07934924215078354,
-0.032372765243053436,
-0.013586361892521381,
-0.06167207658290863,
-0.02966136671602726,
0.01125930156558752,
-0.04710334539413452,
0.01325249020010233,
-0.13308043777942657,
-0.03166798874735832,
-0.07249011844396591,
-0.03646242618560791,
0.18139170110225677,
0.0008431989699602127,
0.02489018253982067,
-0.0740787610411644,
0.032351333647966385,
0.016594283282756805,
-0.12810783088207245,
-0.06389806419610977,
-0.03837756812572479,
-0.05608600750565529,
0.0016061189817264676,
-0.04196035861968994,
-0.06144925206899643,
0.10940535366535187,
0.14352825284004211,
-0.012309943325817585,
0.0202387273311615,
-0.07189576327800751,
0.06052647903561592,
0.03414542227983475,
0.03214697912335396,
-0.1220976784825325,
-0.11984395235776901,
0.032824866473674774,
0.02351100742816925,
0.1045135036110878,
0.0011113585205748677,
-0.07070562243461609,
-0.015548329800367355,
0.007006076164543629,
0.06746502965688705,
0.017366383224725723,
0.07258334755897522,
-0.021677106618881226,
-0.02888043038547039,
0.21923337876796722,
-0.06252351403236389,
-0.12760132551193237,
-0.058974333107471466,
-0.051818206906318665,
0.021114643663167953,
-0.003697590669617057,
0.0516151525080204,
0.03194557875394821,
0.058858439326286316,
-0.036046046763658524,
-0.07819225639104843,
-0.04322899505496025,
-0.011702789925038815,
-0.0019714836962521076,
-0.06494898349046707,
0.011101833544671535,
-0.12355303764343262,
-0.25488847494125366,
-0.048665743321180344,
-0.02292582206428051,
-0.047644175589084625,
-0.05299492925405502,
-0.05662882700562477,
-0.12357205897569656,
0.011308201588690281,
-0.07577740401029587,
0.02766197733581066,
-0.04814279451966286,
0.043984293937683105,
0.05509934946894646,
0.10882874578237534,
-0.02800745703279972,
0.060882095247507095,
-0.08959559351205826,
0.05807576701045036,
-0.14321832358837128,
0.03740234300494194,
-0.09647149592638016,
0.02431005984544754,
-0.17781931161880493,
-0.03355879709124565,
-0.002125505357980728,
0.04172654077410698,
0.019195711240172386,
0.16312888264656067,
-0.137015238404274,
-0.05154787003993988,
0.20243598520755768,
-0.082832470536232,
-0.13880009949207306,
0.12486453354358673,
0.040553055703639984,
-0.013164378702640533,
0.014239413663744926,
0.11376949399709702,
0.002949048997834325,
-0.06637337058782578,
-0.033096715807914734,
0.04399973899126053,
-0.07044905424118042,
-0.05065125972032547,
0.14934615790843964,
0.018345149233937263,
-0.06166287884116173,
0.03782555088400841,
-0.007032555527985096,
0.06336257606744766,
-0.09707049280405045,
-0.049024879932403564,
0.008935954421758652,
-0.10723687708377838,
0.12313024699687958,
-0.07395637780427933,
0.06081806495785713,
-0.010973569937050343,
-0.06337805837392807,
-0.048480190336704254,
0.10934767872095108,
0.04324694350361824,
0.0031014203559607267,
-0.1528017222881317,
0.12056848406791687,
-0.11529581993818283,
0.03320251405239105,
-0.1572767198085785,
-0.07101312279701233,
-0.013607898727059364,
-0.005650118459016085,
0.008650310337543488,
-0.004386341664940119,
0.03067220188677311,
0.0033827531151473522,
-0.036069002002477646,
-0.02736354060471058,
0.0740608349442482,
-0.007863380946218967,
-0.05161621421575546,
-0.10791520774364471,
-0.011181915178894997,
-0.06349495053291321,
0.11617659777402878,
-0.1240982934832573,
0.01876118965446949,
-0.02772233448922634,
0.17068561911582947,
0.04698384925723076,
0.029959918931126595,
0.07775465399026871,
0.040696363896131516,
-0.0014774241717532277,
-0.0007707338081672788,
0.09649082273244858,
-0.004407590255141258,
-0.11650361865758896,
0.13140225410461426,
-0.018592141568660736,
0.010148418135941029,
0.10026894509792328,
0.05391282960772514,
-0.033534396439790726,
-0.049967072904109955,
-0.04641878232359886,
0.013892396353185177,
0.05245915800333023,
0.005172227043658495,
0.08063861727714539,
0.05288435518741608,
0.12606869637966156,
-0.04749935120344162,
0.0003015427791979164,
0.04957261309027672,
0.005262000951915979,
-0.08676382154226303,
0.08641177415847778,
0.05713740363717079,
-0.11062565445899963,
0.06846389174461365,
0.15510034561157227,
-0.0464928075671196,
0.14523421227931976,
0.002996434923261404,
-0.0239267535507679,
-0.004170498810708523,
-0.011499820277094841,
0.0005224032793194056,
0.05016344040632248,
-0.04599461331963539,
-0.01803766004741192,
0.07055026292800903,
0.039084576070308685,
0.023337310180068016,
-0.07317907363176346,
-0.025663645938038826,
-0.05798523873090744,
-0.014290349557995796,
0.030431833118200302,
0.15577544271945953,
-0.03741183504462242,
0.1028655394911766,
0.06519390642642975,
-0.017182089388370514,
0.028384627774357796,
-0.03501700609922409,
-0.059604212641716,
0.1603555977344513,
-0.07428645342588425,
-0.14290380477905273,
-0.11259102821350098,
-0.0017923369305208325,
-0.0784667432308197,
-0.032938409596681595,
0.05507393553853035,
-0.10644552111625671,
-0.010021801106631756,
-0.06995139271020889,
0.044361863285303116,
-0.00823255069553852,
-0.02644997276365757,
0.06814856082201004,
0.015704331919550896,
-0.02144988253712654,
-0.08728998899459839,
0.016504686325788498,
-0.014445879496634007,
-0.11084025353193283,
0.07468158006668091,
-0.01719803176820278,
0.15519224107265472,
0.017858359962701797,
0.01871587708592415,
-0.009445125237107277,
-0.027434073388576508,
0.15486547350883484,
-0.0798821821808815,
-0.0010500969365239143,
0.2854979634284973,
0.02547531947493553,
0.003684538183733821,
0.09816130995750427,
-0.0200445968657732,
-0.08121763914823532,
0.058158744126558304,
0.06135108321905136,
-0.058585140854120255,
-0.20696179568767548,
-0.12418562918901443,
-0.050804294645786285,
0.0290386825799942,
-0.011723327450454235,
0.03794851526618004,
-0.0010658098617568612,
0.08827652782201767,
-0.06144220754504204,
-0.0128548052161932,
-0.06569967418909073,
0.07089819014072418,
0.0636744275689125,
-0.026693889871239662,
0.05878043547272682,
-0.042451344430446625,
-0.05268316715955734,
0.057160258293151855,
0.07968835532665253,
0.12192588299512863,
0.0450335368514061,
0.049874696880578995,
0.0748530924320221,
0.12549494206905365,
0.05205174535512924,
-0.01834350824356079,
-0.025272341445088387,
-0.015391445718705654,
-0.03300325199961662,
-0.07063064724206924,
-0.025138793513178825,
0.06280897557735443,
-0.04812692105770111,
-0.06525994837284088,
-0.017893606796860695,
0.02395499125123024,
0.0305617842823267,
0.21793736517429352,
0.04628128558397293,
-0.23862622678279877,
-0.06857533752918243,
0.08674654364585876,
0.0211092047393322,
-0.04495706036686897,
0.12261878699064255,
0.006469625048339367,
-0.12412425875663757,
0.17107613384723663,
-0.08305902034044266,
0.07894375920295715,
-0.08517179638147354,
0.015421257354319096,
-0.00037741579581052065,
0.07315336912870407,
0.01939302496612072,
0.06622110307216644,
-0.20554836094379425,
0.2213905304670334,
0.008893715217709541,
0.024082044139504433,
-0.066322460770607,
0.011637994088232517,
0.030491003766655922,
0.06429595500230789,
0.1979820877313614,
-0.032749440521001816,
-0.11411061882972717,
-0.102141834795475,
-0.04801665246486664,
0.06205511465668678,
0.10128752887248993,
-0.07885448634624481,
0.0903891995549202,
-0.0009356657974421978,
-0.022981472313404083,
-0.03600795194506645,
-0.046647947281599045,
-0.14812356233596802,
-0.058600109070539474,
0.03256263956427574,
0.09932506829500198,
0.16296960413455963,
-0.047496143728494644,
-0.04970648139715195,
-0.09046750515699387,
0.06653271615505219,
-0.08447825163602829,
-0.04215657711029053,
-0.13679994642734528,
0.003633663523942232,
0.1372309774160385,
-0.07069666683673859,
-0.04814724996685982,
-0.02598973736166954,
0.09924527257680893,
0.007984273135662079,
-0.11053680628538132,
0.08462174236774445,
-0.06757277995347977,
-0.1807829737663269,
-0.025962939485907555,
0.09069477021694183,
0.040879540145397186,
0.02749253809452057,
-0.059745755046606064,
0.0320703350007534,
0.006955568678677082,
-0.11307244747877121,
0.10437259823083878,
0.10740864276885986,
0.047215625643730164,
0.07541432231664658,
-0.09464499354362488,
-0.03681022673845291,
-0.029922468587756157,
-0.056680064648389816,
0.11107664555311203,
0.281689316034317,
-0.032697368413209915,
0.11104309558868408,
0.13573741912841797,
-0.016285313293337822,
-0.2573849856853485,
-0.055538542568683624,
0.002378969918936491,
0.037322841584682465,
-0.028361205011606216,
-0.16493898630142212,
0.09598084539175034,
0.04476280137896538,
-0.05742974206805229,
-0.0042858426459133625,
-0.2538851797580719,
-0.1295146346092224,
0.059686459600925446,
0.026141220703721046,
0.04806036129593849,
-0.13695243000984192,
-0.054808083921670914,
-0.09900012612342834,
-0.19716888666152954,
0.06206179037690163,
-0.12213511019945145,
0.0890854075551033,
-0.06457506865262985,
0.06015820801258087,
-0.03456345200538635,
-0.017887722700834274,
0.09369110316038132,
-0.012159009464085102,
0.02634940668940544,
-0.0674256831407547,
-0.054586850106716156,
0.016828935593366623,
-0.030076581984758377,
0.11131420731544495,
-0.044319216161966324,
0.09139746427536011,
-0.14308656752109528,
-0.004663479048758745,
-0.09670340269804001,
0.09910695254802704,
0.037030018866062164,
-0.030411282554268837,
-0.05760875344276428,
0.0825347825884819,
0.008013595826923847,
0.005743440240621567,
0.24066463112831116,
-0.06187724694609642,
0.09496358782052994,
0.03382927179336548,
0.10334312170743942,
0.020306909456849098,
0.09697218984365463,
-0.04733328893780708,
-0.015417298302054405,
0.07899098098278046,
-0.13202673196792603,
-0.03398573771119118,
0.09501182287931442,
0.0340476855635643,
0.056367695331573486,
0.0062138414941728115,
-0.04629660025238991,
0.12686915695667267,
0.06072597578167915,
-0.10213377326726913,
-0.16490405797958374,
-0.05185091868042946,
0.1237209141254425,
-0.08401007205247879,
0.1218574047088623,
0.18101152777671814,
-0.06833789497613907,
-0.0335332527756691,
-0.0580279566347599,
0.05869576707482338,
-0.07671384513378143,
0.0591103620827198,
0.003591216402128339,
0.031170228496193886,
-0.10017284005880356,
0.10152929276227951,
0.02325151488184929,
-0.06769721210002899,
0.03896636888384819,
0.09269976615905762,
-0.14442390203475952,
-0.06096713989973068,
-0.016041558235883713,
0.18594790995121002,
-0.11355312168598175,
-0.06578322499990463,
-0.06375367194414139,
-0.06913294643163681,
0.04183513671159744,
0.16770997643470764,
0.05580536648631096,
-0.04619772359728813,
-0.010769004933536053,
-0.03926658630371094,
-0.07425303012132645,
0.046317897737026215,
-0.025788715109229088,
0.03680766746401787,
-0.09009922295808792,
0.007451867219060659,
-0.03167039528489113,
0.12895452976226807,
-0.08283862471580505,
-0.013666377402842045,
-0.08107108622789383,
-0.027058495208621025,
-0.15585440397262573,
-0.03438577428460121,
-0.13570040464401245,
0.006290079094469547,
-0.02220059372484684,
-0.08360692113637924,
-0.0820644274353981,
-0.0063078501261770725,
-0.06391174346208572,
-0.017559276893734932,
-0.06605182588100433,
0.10450044274330139,
-0.08366773277521133,
0.02579440549015999,
0.011828710325062275,
-0.020461728796362877,
0.1216362938284874,
0.007352369371801615,
0.020910978317260742,
0.05768643692135811,
-0.10651037096977234,
-0.024176741018891335,
0.01689152605831623,
0.01886533573269844,
0.042571038007736206,
-0.03345012664794922,
-0.015224035829305649,
0.006598723586648703,
0.003169650910422206,
0.01575806364417076,
0.03317086771130562,
-0.09077909588813782,
-0.06374680995941162,
-0.07311950623989105,
0.031933803111314774,
-0.09218374639749527,
0.029987484216690063,
0.10812534391880035,
-0.01018516719341278,
0.1081148087978363,
-0.0771825984120369,
0.0166787076741457,
-0.11366256326436996,
0.001073797233402729,
-0.01295283529907465,
-0.1229388639330864,
-0.05103899911046028,
-0.014274666085839272,
0.06770142912864685,
-0.03976598381996155,
0.07953333854675293,
-0.03182590380311012,
0.034020427614450455,
0.07563310116529465,
-0.053648896515369415,
-0.04387437179684639,
0.03942085802555084,
0.09818845987319946,
0.140152707695961,
0.01984376460313797,
-0.008346395567059517,
0.06432602554559708,
0.016188014298677444,
0.024584326893091202,
0.10541700571775436,
0.14355415105819702,
0.02744937501847744,
0.08778104931116104,
0.08287361264228821,
-0.10126600414514542,
-0.060878828167915344,
0.0698087140917778,
-0.07439480721950531,
0.0651993602514267,
-0.011301684193313122,
0.09550222009420395,
0.15105953812599182,
-0.059749290347099304,
0.008166479878127575,
-0.03210633993148804,
-0.013755706138908863,
-0.08251378685235977,
-0.10237609595060349,
-0.09644874185323715,
-0.1216600313782692,
0.019720571115612984,
-0.11552093923091888,
0.010330310091376305,
0.0887988805770874,
0.0011121444404125214,
-0.01319889910519123,
0.1328858882188797,
-0.0005651902756653726,
-0.04206082224845886,
0.09506561607122421,
-0.06615747511386871,
0.04854093864560127,
-0.009302948601543903,
-0.06728996336460114,
0.08191931247711182,
-0.0498371422290802,
-0.0123357018455863,
0.04676960036158562,
-0.08723032474517822,
0.053717080503702164,
-0.04436197504401207,
-0.12536153197288513,
0.032661084085702896,
0.06989387422800064,
0.060692522674798965,
0.05051101744174957,
0.03619831055402756,
-0.019706066697835922,
0.016802288591861725,
0.2259465605020523,
-0.03708377107977867,
0.034143101423978806,
-0.14000794291496277,
0.1742943376302719,
0.007993541657924652,
-0.002978722332045436,
-0.03688129037618637,
-0.072647824883461,
-0.06456505507230759,
0.2243935912847519,
0.14573878049850464,
-0.06971296668052673,
-0.017850887030363083,
-0.0015651379944756627,
0.0077977594919502735,
-0.03147359937429428,
0.04405625909566879,
0.09541039913892746,
0.15550915896892548,
-0.059740595519542694,
-0.003818939905613661,
0.02834804728627205,
0.016724668443202972,
0.005648507736623287,
0.029378466308116913,
0.010648283176124096,
-0.05106063932180405,
-0.005347209982573986,
0.1256314069032669,
-0.15459273755550385,
-0.12482676655054092,
-0.056090518832206726,
-0.09483230859041214,
-0.13598470389842987,
-0.04571179300546646,
0.0035955028142780066,
0.0879305899143219,
0.06815404444932938,
-0.009792059659957886,
-0.023231932893395424,
0.06819437444210052,
-0.016814591363072395,
-0.017506800591945648,
-0.050979819148778915,
0.036383237689733505,
-0.0326470285654068,
0.1331106722354889,
-0.04206769913434982,
0.010204498656094074,
0.12148689478635788,
-0.034936241805553436,
-0.11642005294561386,
0.07805478572845459,
0.028658922761678696,
-0.03823452442884445,
0.11170180886983871,
0.18938209116458893,
-0.030001003295183182,
0.16919080913066864,
0.09027889370918274,
-0.028991909697651863,
0.04029101878404617,
-0.023115985095500946,
-0.013307307846844196,
-0.07992380857467651,
0.08042485266923904,
-0.09791980683803558,
0.14858806133270264,
0.11415347456932068,
-0.054649125784635544,
0.017352428287267685,
-0.009354892186820507,
0.09332284331321716,
-0.0049706134013831615,
0.08946248888969421,
-0.029591670259833336,
-0.15186966955661774,
-0.04912971705198288,
0.031562644988298416,
-0.012393993325531483,
-0.3033968508243561,
0.027258744463324547,
-0.07638085633516312,
0.026304330676794052,
-0.03445543721318245,
0.023758210241794586,
0.09854207932949066,
-0.011937295086681843,
-0.04548003524541855,
-0.1791478991508484,
0.029197080060839653,
0.1235969141125679,
-0.18831150233745575,
-0.10197946429252625
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | question-answering | Poulami/muril-base-cased-finetuned-QA-SQuADv2 | [
"transformers",
"safetensors",
"bert",
"question-answering",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:32:37+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #question-answering #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #question-answering #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
39,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #question-answering #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05259126052260399,
0.2028854638338089,
-0.0044862194918096066,
0.02425798401236534,
0.10576079040765762,
0.0068638501688838005,
0.06693174690008163,
0.1155666932463646,
0.002374236471951008,
0.12093950808048248,
0.028730813413858414,
0.0795159563422203,
0.11937985569238663,
0.150146484375,
0.0030903066508471966,
-0.24085353314876556,
0.05294279009103775,
-0.09337051957845688,
0.0006256133783608675,
0.1166968047618866,
0.13365985453128815,
-0.108995720744133,
0.08927806466817856,
-0.010429061949253082,
-0.015317060984671116,
-0.016092488542199135,
-0.07380294054746628,
-0.06513319909572601,
0.057247381657361984,
0.07617297768592834,
0.0721721425652504,
0.011596176773309708,
0.076545849442482,
-0.28416764736175537,
0.016033722087740898,
0.08059533685445786,
0.0017565848538652062,
0.06882699579000473,
0.08010933548212051,
-0.06467695534229279,
0.12191252410411835,
-0.06272949278354645,
0.13958317041397095,
0.07744473963975906,
-0.08956745266914368,
-0.19547611474990845,
-0.06588393449783325,
0.06942369788885117,
0.13162168860435486,
0.05067950859665871,
-0.03308458253741264,
0.14173483848571777,
-0.10055480897426605,
0.008528374135494232,
0.12415886670351028,
-0.07726802676916122,
-0.05673770606517792,
0.03650301694869995,
0.10443621873855591,
0.0974099189043045,
-0.11856937408447266,
-0.0074502998031675816,
0.02983100153505802,
0.02229507826268673,
0.08182867616415024,
0.01513586938381195,
0.14877605438232422,
0.042825911194086075,
-0.14221304655075073,
-0.0532369539141655,
0.09163448959589005,
0.03891806676983833,
-0.04990224912762642,
-0.23594164848327637,
-0.026718372479081154,
-0.005824713502079248,
-0.03001447580754757,
-0.04385225474834442,
0.052589140832424164,
-0.03746987506747246,
0.07201015204191208,
-0.00779232382774353,
-0.07770378142595291,
-0.03234218806028366,
0.052429281175136566,
0.06348426640033722,
0.02040219120681286,
-0.00640451954677701,
0.018108926713466644,
0.11223319917917252,
0.0766364336013794,
-0.129673033952713,
-0.07402467727661133,
-0.07489833235740662,
-0.09804271161556244,
-0.04336697235703468,
0.03731705620884895,
0.07777538150548935,
0.0488663986325264,
0.20594625174999237,
-0.017311906442046165,
0.04591337591409683,
0.04501218721270561,
0.00415299367159605,
0.07107240706682205,
0.10608655214309692,
-0.06711014360189438,
-0.1433931291103363,
-0.061982374638319016,
0.08940984308719635,
-0.011836115270853043,
-0.03700637072324753,
-0.046884775161743164,
0.04537593573331833,
0.03191437944769859,
0.11517442017793655,
0.08723600953817368,
-0.0026339609175920486,
-0.06073928251862526,
-0.04193990305066109,
0.22032427787780762,
-0.1429872363805771,
0.04091168940067291,
0.0095632029697299,
-0.04190882295370102,
-0.010128729976713657,
0.0027068995404988527,
0.01748555712401867,
-0.03194664046168327,
0.10472383350133896,
-0.07686623185873032,
-0.03826445713639259,
-0.11045826226472855,
-0.06568188220262527,
0.02656155452132225,
0.002025776309892535,
-0.019147520884871483,
-0.04228195920586586,
-0.1093553751707077,
-0.05163319781422615,
0.07104253768920898,
-0.07569357752799988,
-0.05701810121536255,
0.010968777351081371,
-0.05083809792995453,
0.0019326701294630766,
0.00003669456782517955,
0.11747599393129349,
-0.03144557401537895,
0.027406498789787292,
-0.04884810373187065,
0.06846722960472107,
0.11113353818655014,
0.037212807685136795,
-0.08160411566495895,
0.07096786797046661,
-0.22421935200691223,
0.10047981888055801,
-0.09533138573169708,
0.02410157397389412,
-0.14949576556682587,
-0.04325840622186661,
0.029235901311039925,
0.029889225959777832,
-0.010648856870830059,
0.12825602293014526,
-0.1955174207687378,
-0.03177010640501976,
0.15263482928276062,
-0.11147930473089218,
-0.09708426892757416,
0.06840939819812775,
-0.05585639923810959,
0.11500564217567444,
0.047934968024492264,
-0.019564000889658928,
0.0642162337899208,
-0.1284250169992447,
-0.0425708033144474,
-0.022211723029613495,
-0.0038072247989475727,
0.1497415006160736,
0.06880437582731247,
-0.05901612341403961,
0.0664527639746666,
0.01987437717616558,
-0.040110185742378235,
-0.0367138609290123,
-0.037162408232688904,
-0.09525077044963837,
0.010196850635111332,
-0.07203976064920425,
0.036134399473667145,
-0.019966883584856987,
-0.08927342295646667,
-0.03179198503494263,
-0.17647308111190796,
0.0388604961335659,
0.07913081347942352,
0.009538567624986172,
-0.0228617861866951,
-0.09321427345275879,
0.02225036732852459,
-0.009768214076757431,
-0.020878251641988754,
-0.1617123931646347,
-0.049427129328250885,
0.04355454072356224,
-0.19472132623195648,
0.022555366158485413,
-0.03160899132490158,
0.05118642747402191,
0.029774360358715057,
-0.04335848614573479,
-0.002816336927935481,
0.005193492863327265,
0.017469679936766624,
-0.027759965509176254,
-0.20299731194972992,
-0.02891785092651844,
-0.02918923646211624,
0.13617254793643951,
-0.2220425307750702,
0.02758612483739853,
0.07672455906867981,
0.1353377252817154,
0.0010315107647329569,
-0.042468901723623276,
0.012285644188523293,
-0.05271144583821297,
-0.05070718750357628,
-0.062143102288246155,
-0.003471375210210681,
-0.03387059271335602,
-0.04454231262207031,
0.0628012865781784,
-0.19397759437561035,
-0.03897572308778763,
0.10589398443698883,
0.10167904943227768,
-0.14617563784122467,
-0.031377632170915604,
-0.04175719991326332,
-0.061136167496442795,
-0.09209086000919342,
-0.06178439036011696,
0.1394909918308258,
0.04897880181670189,
0.04573993757367134,
-0.0880829468369484,
-0.06276343762874603,
0.00770267890766263,
-0.0005318898474797606,
-0.03498470038175583,
0.09247587621212006,
0.08966474235057831,
-0.10926453024148941,
0.09032607078552246,
0.08584415167570114,
0.06689407676458359,
0.1032552719116211,
-0.001143817906267941,
-0.10756317526102066,
-0.02851158007979393,
0.010317605920135975,
0.0138178039342165,
0.14097030460834503,
-0.0487520769238472,
0.049268294125795364,
0.05768076702952385,
-0.027659596875309944,
0.016742954030632973,
-0.103117935359478,
0.02645859867334366,
0.042582202702760696,
-0.01090834941715002,
0.011281476356089115,
-0.03452678769826889,
0.027541857212781906,
0.09036476910114288,
0.036810778081417084,
0.028066303580999374,
0.003723195753991604,
-0.036288049072027206,
-0.10548990219831467,
0.17550140619277954,
-0.08578408509492874,
-0.28400298953056335,
-0.1329312026500702,
-0.007224106229841709,
0.050459496676921844,
-0.02486414648592472,
0.015295456163585186,
-0.05430449917912483,
-0.11374207586050034,
-0.10182642191648483,
0.010211742483079433,
0.04778311029076576,
-0.07805939763784409,
-0.07185705006122589,
0.04724802076816559,
0.03216748312115669,
-0.1409207433462143,
0.023733584210276604,
0.04673764482140541,
-0.042380258440971375,
-0.011882458813488483,
0.06965228915214539,
0.10261393338441849,
0.1735156923532486,
-0.0012796181254088879,
-0.018994955345988274,
0.02287985570728779,
0.24996212124824524,
-0.1460387110710144,
0.10694320499897003,
0.15233273804187775,
-0.06607743352651596,
0.10274875164031982,
0.19897888600826263,
0.0273613091558218,
-0.07705751806497574,
0.0358937531709671,
0.04155202955007553,
-0.05248534306883812,
-0.2364395558834076,
-0.05910567194223404,
0.001085618045181036,
-0.06965078413486481,
0.08993931114673615,
0.0902039185166359,
0.11398119479417801,
0.04661845788359642,
-0.08674395084381104,
-0.07518473267555237,
0.020830264315009117,
0.10745775699615479,
-0.030370062217116356,
0.006791229825466871,
0.09171789139509201,
-0.04242653772234917,
-0.007607475854456425,
0.10837430506944656,
0.011178508400917053,
0.19777020812034607,
0.03209175914525986,
0.15326212346553802,
0.07447052001953125,
0.03400443494319916,
0.027105089277029037,
0.02122676372528076,
0.022011037915945053,
0.008528202772140503,
-0.017223529517650604,
-0.08776357024908066,
0.03130582347512245,
0.13481217622756958,
0.07551243901252747,
0.026613935828208923,
0.023921193554997444,
-0.04104981943964958,
0.06177108362317085,
0.164777934551239,
0.007725866511464119,
-0.2068057656288147,
-0.03830723464488983,
0.09051752090454102,
-0.08040063828229904,
-0.12359555810689926,
-0.02303191088140011,
0.04106256365776062,
-0.18250946700572968,
0.038382045924663544,
-0.017931431531906128,
0.11581720411777496,
-0.11333613097667694,
-0.02766553685069084,
0.04014914482831955,
0.07990425080060959,
-0.03312310576438904,
0.0804636999964714,
-0.17019516229629517,
0.12019498646259308,
0.012228374369442463,
0.05620236322283745,
-0.11772795766592026,
0.09470738470554352,
0.01082859467715025,
-0.012395285069942474,
0.17020393908023834,
-0.0009680894436314702,
-0.07143643498420715,
-0.06312961876392365,
-0.07824177294969559,
-0.019154205918312073,
0.09248737245798111,
-0.11479625850915909,
0.08222582191228867,
-0.01338129211217165,
-0.0378129743039608,
0.0046119531616568565,
-0.10743723064661026,
-0.12412271648645401,
-0.1919620931148529,
0.058793265372514725,
-0.11608237028121948,
-0.00009404727461514995,
-0.09927497804164886,
-0.0562993660569191,
-0.047756634652614594,
0.1977493017911911,
-0.13937753438949585,
-0.09716302901506424,
-0.1498957872390747,
-0.09690418094396591,
0.16912008821964264,
-0.04737471044063568,
0.08579006046056747,
-0.0020281081087887287,
0.22024127840995789,
0.0072343116626143456,
-0.008195682428777218,
0.07956323027610779,
-0.08450262248516083,
-0.18208858370780945,
-0.07643314450979233,
0.12372209876775742,
0.12876582145690918,
0.053298406302928925,
-0.01190928928554058,
0.022812241688370705,
-0.03731144964694977,
-0.11800242960453033,
0.005281681660562754,
0.11922437697649002,
0.06110478565096855,
0.039771661162376404,
-0.0013343535829335451,
-0.09944376349449158,
-0.07634468376636505,
-0.03782802075147629,
0.02442140504717827,
0.19300124049186707,
-0.08433309197425842,
0.15842226147651672,
0.134726420044899,
-0.05046136677265167,
-0.21320460736751556,
0.03127184882760048,
0.04307406395673752,
0.0005565352621488273,
0.05940066650509834,
-0.17698396742343903,
0.0836062878370285,
0.022285189479589462,
-0.0521271750330925,
0.14751119911670685,
-0.16639907658100128,
-0.1545284390449524,
0.08442960679531097,
0.051656633615493774,
-0.21620962023735046,
-0.12632134556770325,
-0.08712536841630936,
-0.06352436542510986,
-0.1410045474767685,
0.08878432959318161,
0.011481193825602531,
-0.002163688186556101,
0.04926665499806404,
0.03752978518605232,
0.019776154309511185,
-0.050277549773454666,
0.2175254374742508,
-0.001740340725518763,
0.03035682998597622,
-0.07569415867328644,
-0.1040424332022667,
0.06715744733810425,
-0.048464689403772354,
0.08361992985010147,
-0.02904176339507103,
0.00727926567196846,
-0.08175947517156601,
-0.05732819065451622,
-0.0551212914288044,
0.030917048454284668,
-0.08432964235544205,
-0.10519272089004517,
-0.06682265549898148,
0.09636414796113968,
0.09235509485006332,
-0.033162232488393784,
-0.03963904082775116,
-0.09284978359937668,
0.04546565189957619,
0.20358207821846008,
0.1786501407623291,
0.05575625225901604,
-0.09733896702528,
0.004446255974471569,
-0.02058122307062149,
0.041350848972797394,
-0.21201425790786743,
0.049836933612823486,
0.04702357202768326,
0.02481400966644287,
0.12029092758893967,
-0.019803699105978012,
-0.16318121552467346,
-0.05019678175449371,
0.05431952700018883,
-0.03671557083725929,
-0.21271918714046478,
-0.012301293201744556,
0.05155809968709946,
-0.19053103029727936,
-0.06799828261137009,
0.015656087547540665,
-0.009748312644660473,
-0.027901990339159966,
0.01566028594970703,
0.06412044167518616,
0.028648724779486656,
0.0990493968129158,
0.06228994205594063,
0.10199911147356033,
-0.11336737871170044,
0.07902315258979797,
0.08970371633768082,
-0.08288145065307617,
0.010017158463597298,
0.07221437245607376,
-0.05355165898799896,
-0.024539576843380928,
0.0245730672031641,
0.06422583013772964,
-0.0015890412032604218,
-0.056649502366781235,
-0.019970232620835304,
-0.10758348554372787,
0.06470255553722382,
0.12866966426372528,
0.039797183126211166,
-0.0097538772970438,
0.051341574639081955,
0.02343105897307396,
-0.07750807702541351,
0.11568180471658707,
0.03222927078604698,
0.03643401339650154,
-0.05755656212568283,
-0.024085741490125656,
0.041584983468055725,
0.007216460537165403,
-0.020499495789408684,
-0.025817055255174637,
-0.05941548943519592,
-0.012329903431236744,
-0.1958753764629364,
0.013851859606802464,
-0.07513978332281113,
0.006235042586922646,
0.014809658750891685,
-0.04453662782907486,
-0.019246714189648628,
0.010231224820017815,
-0.08321903645992279,
-0.04743500053882599,
-0.002234705025330186,
0.09950006753206253,
-0.1428065150976181,
0.008886833675205708,
0.09462965279817581,
-0.11724920570850372,
0.0701248049736023,
-0.023760873824357986,
-0.01835380494594574,
0.002579940715804696,
-0.12669740617275238,
0.045131903141736984,
-0.007041062228381634,
0.020651215687394142,
0.04380253329873085,
-0.17144426703453064,
0.002618284896016121,
-0.04253067821264267,
-0.04294731095433235,
-0.0120546268299222,
-0.08081796765327454,
-0.11252862215042114,
0.11146305501461029,
0.0013757607666775584,
-0.08535416424274445,
-0.010495108552277088,
0.04560060799121857,
0.10778488218784332,
-0.03788670152425766,
0.1243036538362503,
0.005103552248328924,
0.06758087128400803,
-0.18262599408626556,
-0.024756142869591713,
-0.014731800183653831,
0.008348725736141205,
0.024925386533141136,
-0.01479937694966793,
0.042617179453372955,
-0.015035021118819714,
0.26043418049812317,
-0.02653587982058525,
0.074964240193367,
0.06469603627920151,
0.042822424322366714,
0.0198125708848238,
0.08178427815437317,
0.06811564415693283,
0.016192639246582985,
0.004397526383399963,
0.02518327906727791,
-0.03391282260417938,
-0.02226550132036209,
-0.15770775079727173,
0.08034674823284149,
0.15029585361480713,
0.09046841412782669,
0.0038341283798217773,
0.06518141180276871,
-0.10338674485683441,
-0.10085316747426987,
0.08406507968902588,
-0.03721795231103897,
-0.004446348641067743,
-0.06144845485687256,
0.15846431255340576,
0.15133251249790192,
-0.17548325657844543,
0.08107186108827591,
-0.04381648451089859,
-0.049447815865278244,
-0.11742796003818512,
-0.1601942777633667,
-0.06760376691818237,
-0.031142009422183037,
-0.0010461981873959303,
-0.057294074445962906,
0.06657780706882477,
0.11034563183784485,
0.00035543262492865324,
-0.0012925605988129973,
0.0943048745393753,
-0.024420278146862984,
-0.016321783885359764,
0.033111512660980225,
0.0479465015232563,
0.038273297250270844,
-0.04731360450387001,
0.02323073334991932,
0.005872312933206558,
0.04081957787275314,
0.059196196496486664,
0.024727536365389824,
-0.041051145642995834,
0.013183102011680603,
-0.014876319095492363,
-0.1053827702999115,
0.0206439346075058,
-0.027507297694683075,
-0.07129579037427902,
0.13602343201637268,
0.02897627092897892,
0.01998012699186802,
-0.03161655366420746,
0.2050580531358719,
-0.07159226387739182,
-0.07268409430980682,
-0.14224226772785187,
0.1074400320649147,
-0.04027022793889046,
0.06320521980524063,
0.05849554017186165,
-0.1132800355553627,
-0.0029967257287353277,
0.13321146368980408,
0.13381916284561157,
-0.03547452762722969,
0.005371219478547573,
0.03219069167971611,
0.008148705586791039,
-0.048041801899671555,
0.04534820839762688,
0.03209371492266655,
0.15637032687664032,
-0.07269488275051117,
0.075834721326828,
0.0010619045933708549,
-0.08687957376241684,
-0.03515656664967537,
0.14211764931678772,
0.00892868172377348,
0.027962028980255127,
-0.06807447224855423,
0.10675215721130371,
-0.07595851272344589,
-0.24105240404605865,
0.04082513973116875,
-0.07905600219964981,
-0.15839076042175293,
-0.014650527387857437,
0.025225967168807983,
-0.009219054132699966,
0.02357676811516285,
0.0644829124212265,
-0.06328418105840683,
0.16023315489292145,
0.03488091379404068,
-0.08454935252666473,
-0.05776360258460045,
0.08267566561698914,
-0.09477023035287857,
0.2949345111846924,
0.014565679244697094,
0.03338702395558357,
0.10453473776578903,
-0.023210318759083748,
-0.1319405734539032,
0.028264416381716728,
0.10424477607011795,
-0.1026621162891388,
0.06628262251615524,
0.18118372559547424,
-0.008201897144317627,
0.10254274308681488,
0.07821037620306015,
-0.06196645274758339,
0.05818212404847145,
-0.08231779932975769,
-0.06688329577445984,
-0.09811308234930038,
0.0639098510146141,
-0.06666210293769836,
0.14476773142814636,
0.12259301543235779,
-0.0448145866394043,
-0.004128707572817802,
-0.02726835571229458,
0.0432763397693634,
0.011227069422602654,
0.12515388429164886,
0.010173868387937546,
-0.1640346348285675,
0.028346290811896324,
0.0038582251872867346,
0.1055222600698471,
-0.22144944965839386,
-0.0828825980424881,
0.05016852542757988,
-0.027500580996274948,
-0.054417576640844345,
0.1055392250418663,
0.06477102637290955,
0.049571000039577484,
-0.04528665170073509,
-0.060297247022390366,
-0.010412150993943214,
0.15066713094711304,
-0.11862421035766602,
-0.005452236160635948
] |
null | null | stable-baselines3 |
# **A2C** Agent playing **PandaReachDense-v3**
This is a trained model of a **A2C** agent playing **PandaReachDense-v3**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["PandaReachDense-v3", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "A2C", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "PandaReachDense-v3", "type": "PandaReachDense-v3"}, "metrics": [{"type": "mean_reward", "value": "-0.14 +/- 0.09", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | r0in/a2c-PandaReachDense-v3 | [
"stable-baselines3",
"PandaReachDense-v3",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-06T06:35:50+00:00 | [] | [] | TAGS
#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# A2C Agent playing PandaReachDense-v3
This is a trained model of a A2C agent playing PandaReachDense-v3
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
41,
45,
17
] | [
"passage: TAGS\n#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.028780510649085045,
0.06549051403999329,
-0.004174588713794947,
0.028733979910612106,
0.12748076021671295,
-0.010029550641775131,
0.16130082309246063,
0.07903143763542175,
0.052706290036439896,
-0.055043965578079224,
0.09157051891088486,
-0.079488605260849,
0.04699381813406944,
0.3393711447715759,
0.029525093734264374,
-0.186785027384758,
0.08573613315820694,
0.015584449283778667,
0.018966808915138245,
0.09867662936449051,
0.03466832637786865,
-0.08736564218997955,
0.04568251967430115,
0.03800429776310921,
-0.07686931639909744,
-0.04319252818822861,
-0.03975098207592964,
-0.06744661927223206,
0.10361767560243607,
-0.044310007244348526,
0.1670169234275818,
-0.03489987552165985,
0.10219604521989822,
-0.12577489018440247,
0.031373992562294006,
-0.04813149571418762,
-0.05141052231192589,
0.002818689215928316,
-0.011371237225830555,
0.05937984213232994,
0.04167760908603668,
0.05197896435856819,
0.07366002351045609,
0.04871916025876999,
-0.08704962581396103,
-0.11396265029907227,
-0.006845315918326378,
0.07931416481733322,
0.17974808812141418,
0.04054044932126999,
-0.02474738284945488,
0.09696658700704575,
-0.11350683122873306,
0.01657135598361492,
-0.019304286688566208,
-0.4018571078777313,
0.006876560393720865,
0.15550047159194946,
0.04677277058362961,
0.010903568007051945,
-0.0061170910485088825,
-0.004642391111701727,
0.02805398777127266,
-0.037410516291856766,
0.08670840412378311,
-0.09000635892152786,
0.06153826415538788,
-0.019131680950522423,
-0.04113767296075821,
-0.01751464419066906,
0.2419518232345581,
0.01633240468800068,
-0.08024721592664719,
-0.07922019064426422,
0.009968155063688755,
-0.028026137501001358,
-0.0877801775932312,
-0.06134319305419922,
0.07644549012184143,
0.057131536304950714,
0.10696670413017273,
-0.030399860814213753,
-0.058683689683675766,
-0.04541248828172684,
0.08352918922901154,
-0.03953780233860016,
-0.017566127702593803,
-0.01754307933151722,
-0.06739802658557892,
-0.003707833355292678,
0.015629740431904793,
-0.06615205854177475,
-0.015486059710383415,
-0.044966671615839005,
-0.1556774228811264,
-0.009128551930189133,
-0.0599384643137455,
0.03310214728116989,
0.10073909163475037,
0.13065455853939056,
0.06838785856962204,
0.09685135632753372,
-0.08001106232404709,
0.0389438234269619,
0.06625691801309586,
0.09461154788732529,
-0.044509198516607285,
-0.011874453164637089,
0.14630302786827087,
0.10327376425266266,
0.09657767415046692,
-0.09182082861661911,
-0.12403369694948196,
0.04173071309924126,
0.10965418070554733,
0.03382069617509842,
0.0046537998132407665,
0.04452834278345108,
-0.14144757390022278,
0.023916395381093025,
0.0006972529226914048,
-0.045244041830301285,
-0.03088594414293766,
0.06111180782318115,
-0.04433412477374077,
0.02348744124174118,
-0.012718633748590946,
0.10830001533031464,
0.10152670741081238,
-0.023899899795651436,
-0.052799396216869354,
-0.04201658070087433,
-0.0440504252910614,
-0.05507666990160942,
0.04012975096702576,
0.01289378758519888,
0.04624854028224945,
-0.1184653639793396,
-0.13997629284858704,
0.051258668303489685,
0.019622454419732094,
-0.026321161538362503,
-0.13472233712673187,
-0.09338399767875671,
-0.03747362270951271,
-0.011210841126739979,
0.0030350966844707727,
-0.19588395953178406,
-0.02434816211462021,
-0.03428230062127113,
0.13725687563419342,
0.10810749977827072,
-0.06433141976594925,
-0.06369391083717346,
-0.12834231555461884,
0.06795675307512283,
-0.23485252261161804,
0.038750845938920975,
-0.09932064265012741,
0.12411006540060043,
0.007471752353012562,
0.023616313934326172,
0.1410844624042511,
0.02330038882791996,
0.004575210623443127,
0.1702503114938736,
-0.18833371996879578,
-0.046672217547893524,
0.17527204751968384,
-0.0857074186205864,
-0.17703735828399658,
0.05021136254072189,
-0.02124672941863537,
-0.013779462315142155,
0.06350992619991302,
0.09937554597854614,
-0.01727774553000927,
-0.17061583697795868,
0.02558896690607071,
-0.0014508399181067944,
-0.05959303304553032,
0.021542999893426895,
0.12072649598121643,
0.08040176331996918,
-0.027203790843486786,
-0.0016989230643957853,
-0.15452547371387482,
0.09701786935329437,
-0.023543400689959526,
-0.08447092026472092,
0.022736359387636185,
-0.10411997884511948,
0.10016260296106339,
-0.015677137300372124,
0.10591494292020798,
-0.02265925332903862,
-0.018805475905537605,
-0.032891299575567245,
0.10408006608486176,
-0.0068649593740701675,
0.039593957364559174,
-0.17728297412395477,
0.1326225996017456,
0.02176543138921261,
0.046730607748031616,
-0.10109715908765793,
-0.10202061384916306,
0.06674831360578537,
0.15375585854053497,
0.05606463924050331,
0.03833417221903801,
0.07328703999519348,
0.03443831577897072,
-0.0030986627098172903,
-0.1205538883805275,
-0.12789975106716156,
0.019881807267665863,
0.06068658083677292,
-0.08039596676826477,
-0.05172275751829147,
-0.10460081696510315,
0.21138279139995575,
-0.10705634206533432,
0.012047823518514633,
-0.09333895146846771,
0.010153836570680141,
0.08388294279575348,
0.01348812971264124,
0.08132237941026688,
0.02585482969880104,
-0.04426883906126022,
0.009419471956789494,
0.0882885605096817,
0.044275086373090744,
-0.1379590630531311,
0.03784618154168129,
0.024114131927490234,
0.23272188007831573,
0.15174852311611176,
-0.016499420627951622,
-0.055556558072566986,
0.006534850224852562,
0.03740030899643898,
0.03533044084906578,
0.034956689924001694,
0.06951800733804703,
0.1090264692902565,
0.07713755965232849,
0.1276414394378662,
-0.05066131055355072,
0.17763042449951172,
-0.006530070677399635,
-0.14888496696949005,
0.02993084490299225,
-0.07033783197402954,
0.0941668227314949,
-0.06030277907848358,
0.048379335552453995,
0.05410725995898247,
0.0304675605148077,
0.08504439890384674,
-0.00693494314327836,
0.022639812901616096,
-0.04341154545545578,
0.04943868890404701,
0.06790532171726227,
0.06545940041542053,
0.06452376395463943,
-0.007423467002809048,
0.015456308610737324,
-0.05288444459438324,
-0.0518295019865036,
-0.10519610345363617,
-0.12370408326387405,
0.037892695516347885,
-0.015912096947431564,
-0.04463989660143852,
-0.01629551686346531,
-0.07266248762607574,
0.050321705639362335,
0.05250744894146919,
-0.07199236750602722,
0.028561361134052277,
-0.007090074475854635,
-0.09633425623178482,
0.1130511462688446,
-0.14269201457500458,
-0.31355980038642883,
-0.02000165916979313,
-0.13154496252536774,
-0.02077566273510456,
0.15819574892520905,
-0.057956792414188385,
-0.1681092083454132,
0.03305667266249657,
-0.02401961199939251,
-0.09238096326589584,
0.04225420579314232,
-0.018061356619000435,
0.10221174359321594,
0.0857708528637886,
0.043082691729068756,
0.00862243864685297,
-0.01184127852320671,
-0.03903079405426979,
-0.08788500726222992,
0.07608162611722946,
-0.06721128523349762,
0.1173204705119133,
0.13519366085529327,
0.04123268276453018,
-0.015909500420093536,
-0.02043113484978676,
0.06215733662247658,
0.012027861550450325,
-0.036599598824977875,
0.13453175127506256,
-0.03608042374253273,
-0.00864011887460947,
0.04470202699303627,
0.008029532618820667,
-0.10533943772315979,
0.09432658553123474,
-0.05022074654698372,
-0.06974482536315918,
-0.017500806599855423,
-0.08790571242570877,
-0.09950723499059677,
0.18995612859725952,
0.0490412712097168,
0.007856572046875954,
-0.05151839926838875,
0.036120012402534485,
0.07772433012723923,
0.044773608446121216,
0.007161281071603298,
0.03985898196697235,
-0.005716364365071058,
-0.013170693069696426,
0.05278664082288742,
-0.023887991905212402,
0.009960537776350975,
-0.007844919338822365,
0.13077811896800995,
-0.015673788264393806,
0.10317149013280869,
0.0030158995650708675,
0.008619097992777824,
0.08018261194229126,
0.12394148856401443,
0.08064290136098862,
0.019240466877818108,
-0.11554506421089172,
-0.04732639715075493,
-0.030522609129548073,
-0.18181301653385162,
0.11669926345348358,
0.10738886147737503,
0.05268440023064613,
-0.05564067140221596,
0.22832486033439636,
0.0012100599706172943,
0.10802210867404938,
0.03496129810810089,
-0.17664514482021332,
0.024751557037234306,
0.03574612736701965,
0.050895314663648605,
0.007034227252006531,
0.062039270997047424,
-0.09453237801790237,
-0.1839483082294464,
0.03968557342886925,
0.018860090523958206,
0.05523261800408363,
-0.018427258357405663,
0.018512532114982605,
-0.12044285237789154,
-0.05746040865778923,
0.02161633037030697,
0.02076297253370285,
-0.3029120862483978,
0.06816349923610687,
-0.04133946821093559,
0.07392577081918716,
0.009542034938931465,
0.01343793235719204,
0.06604447960853577,
0.01652485318481922,
0.1375029981136322,
-0.017935138195753098,
0.1707022786140442,
-0.1572514772415161,
-0.16084668040275574,
0.025680551305413246,
-0.059293005615472794,
0.07245437800884247,
0.082563117146492,
0.017692390829324722,
0.0069250138476490974,
-0.00047057756455615163,
0.20794180035591125,
-0.13032017648220062,
-0.0346711240708828,
-0.035274047404527664,
0.019543148577213287,
0.022580156102776527,
-0.03844551369547844,
-0.021310672163963318,
0.06112392246723175,
0.1489492505788803,
0.07546767592430115,
-0.02780069410800934,
-0.04611911624670029,
-0.03938353434205055,
-0.09507237374782562,
-0.044778671115636826,
0.10472412407398224,
-0.07841785997152328,
0.10144548118114471,
-0.07513871043920517,
-0.04432075098156929,
0.11707907915115356,
-0.09250949323177338,
-0.053160861134529114,
-0.07627046853303909,
0.05462219938635826,
0.008296831510961056,
0.13374868035316467,
0.03642493113875389,
0.02114485390484333,
0.10089845955371857,
-0.05001259222626686,
0.08662480860948563,
0.03777577355504036,
-0.03541218861937523,
0.03517242521047592,
-0.05375073477625847,
-0.04829130321741104,
-0.010828596539795399,
0.03814345970749855,
0.24244728684425354,
0.302570104598999,
-0.012830551713705063,
0.1897524893283844,
0.09193363785743713,
0.029696941375732422,
-0.16292639076709747,
-0.1200476586818695,
0.05548451840877533,
0.059938978403806686,
0.06154406815767288,
-0.2788083851337433,
0.057189684361219406,
-0.053967077285051346,
-0.08999616652727127,
-0.06829255819320679,
-0.08560561388731003,
-0.07613074034452438,
0.088682159781456,
0.08794322609901428,
0.09100460261106491,
-0.12551987171173096,
0.015924450010061264,
-0.012671655975282192,
-0.1664767563343048,
0.12128932029008865,
-0.039350032806396484,
0.07007917016744614,
-0.025050386786460876,
-0.06438229978084564,
0.025165842846035957,
-0.02775278501212597,
0.04424511641263962,
-0.1206880658864975,
0.0005293674184940755,
-0.04527926817536354,
-0.03749620169401169,
0.1088484600186348,
0.020565982908010483,
-0.0028168195858597755,
-0.09558401256799698,
-0.011945599690079689,
-0.3103867173194885,
0.01988539844751358,
0.02114551141858101,
-0.039148375391960144,
-0.0012507046340033412,
-0.08678091317415237,
-0.042053963989019394,
0.10508828610181808,
0.03930897265672684,
0.08641290664672852,
0.15335260331630707,
-0.005581455305218697,
-0.021082017570734024,
0.17506572604179382,
0.05701295658946037,
-0.014002309180796146,
0.10069113969802856,
-0.06732672452926636,
-0.06576105207204819,
0.04418903961777687,
-0.1016126498579979,
-0.005435575265437365,
0.005642053205519915,
-0.007821558974683285,
0.07107745110988617,
0.09962856024503708,
-0.03340476378798485,
0.18194207549095154,
0.09798844903707504,
-0.15048468112945557,
0.0030947427731007338,
0.052597809582948685,
-0.032650984823703766,
0.04424609988927841,
-0.04443032294511795,
0.05541829764842987,
-0.07521786540746689,
-0.03790169581770897,
0.02031708136200905,
-0.01010141521692276,
-0.07618512213230133,
0.00011962707503698766,
0.03176301345229149,
0.029956085607409477,
-0.08340912312269211,
0.14036758244037628,
0.016359949484467506,
0.0652431845664978,
0.11902019381523132,
0.019259776920080185,
-0.10460162162780762,
-0.014167122542858124,
-0.02339506521821022,
0.2028627097606659,
-0.007937151938676834,
-0.018536100164055824,
-0.11391238868236542,
-0.12847240269184113,
0.018047582358121872,
-0.10348039865493774,
0.10282431542873383,
-0.052032727748155594,
-0.06570395082235336,
-0.03704213351011276,
-0.05561172217130661,
0.031932998448610306,
0.017090078443288803,
-0.015642894431948662,
-0.16111870110034943,
-0.04170334339141846,
0.06846143305301666,
0.039452772587537766,
-0.06145704537630081,
-0.06289087235927582,
-0.16302458941936493,
0.03506235405802727,
-0.1278870701789856,
0.0010145133128389716,
-0.047339316457509995,
-0.05002537742257118,
-0.05195476487278938,
0.01521157007664442,
-0.0177876316010952,
0.008817745372653008,
-0.05148332938551903,
0.03292781487107277,
0.011250603944063187,
0.0014076961670070887,
-0.06952075660228729,
-0.04419080913066864,
0.032172493636608124,
-0.04430563375353813,
0.0661356970667839,
0.04131564497947693,
-0.005653871223330498,
0.021474739536643028,
-0.07005896419286728,
-0.10248169302940369,
0.10313672572374344,
-0.014939527027308941,
0.050572704523801804,
-0.0603681318461895,
-0.012018447741866112,
0.007195405196398497,
-0.07569561898708344,
-0.007751014549285173,
0.24328774213790894,
-0.010914106853306293,
-0.05394120141863823,
-0.07426224648952484,
-0.036970075219869614,
-0.09100507944822311,
-0.0004900419735349715,
0.1948854625225067,
0.05477539822459221,
0.14600017666816711,
-0.0532439760863781,
0.08785777539014816,
-0.06481330841779709,
-0.01534446980804205,
-0.08259234577417374,
0.030320849269628525,
-0.157977893948555,
-0.08130980283021927,
-0.028043894097208977,
-0.03728124126791954,
0.13441862165927887,
-0.19242097437381744,
0.0032852457370609045,
-0.010904400609433651,
-0.04910553991794586,
0.11381126195192337,
0.0557032972574234,
0.24474471807479858,
0.1050342544913292,
-0.035265225917100906,
0.10503548383712769,
0.12215624749660492,
0.0929517149925232,
-0.03347417712211609,
0.058777112513780594,
-0.05078745633363724,
-0.0868106484413147,
0.09736774861812592,
0.012061800807714462,
0.036776214838027954,
-0.08157306164503098,
0.022900743409991264,
-0.10047483444213867,
0.002025678288191557,
0.02005080319941044,
0.2473200410604477,
0.1967000812292099,
-0.09632564336061478,
-0.012216159142553806,
-0.05708231031894684,
-0.032561756670475006,
-0.04091155156493187,
-0.002459051087498665,
-0.07821618020534515,
-0.21873407065868378,
0.051539067178964615,
-0.0930585265159607,
-0.07632365822792053,
-0.06189138814806938,
-0.04064059257507324,
-0.02870149537920952,
0.046939339488744736,
0.03212931379675865,
0.04136762022972107,
0.05070297420024872,
-0.0371626541018486,
-0.09345480799674988,
0.06879863888025284,
-0.11172787100076675,
-0.042014576494693756,
-0.03408866748213768,
0.014045859687030315,
0.032319605350494385,
-0.07429610192775726,
0.07487598061561584,
-0.012149554677307606,
-0.07710553705692291,
0.036456044763326645,
-0.03482281416654587,
0.02153356932103634,
0.07482071220874786,
0.04184282198548317,
-0.09644174575805664,
0.015602846629917622,
0.18867559731006622,
0.020273970440030098,
0.008802177384495735,
-0.14742465317249298,
0.2000039666891098,
-0.02619965374469757,
0.07266447693109512,
-0.03337041288614273,
-0.015141828916966915,
-0.10115411877632141,
0.19129611551761627,
0.11998134851455688,
-0.24376079440116882,
0.024953339248895645,
-0.12912821769714355,
0.022151969373226166,
-0.13376696407794952,
0.20840151607990265,
0.05465596541762352,
0.10847201198339462,
-0.06020665541291237,
-0.02479162998497486,
-0.1493310034275055,
-0.09408020973205566,
-0.08478302508592606,
-0.0414455346763134,
0.10249399393796921,
0.0031611735466867685,
-0.05072701349854469,
-0.00887248944491148,
-0.1566619724035263,
0.10201162099838257,
-0.048264030367136,
-0.11855816096067429,
-0.0679796114563942,
-0.059141192585229874,
-0.06102965027093887,
0.11088541150093079,
0.11637356877326965,
-0.01684124954044819,
0.024554423987865448,
-0.07280154526233673,
-0.012559473514556885,
0.011003518477082253,
0.005383014678955078,
0.0626269057393074,
-0.04783647879958153,
0.1594477891921997,
-0.021524829789996147,
0.0008918871753849089,
0.04285505786538124,
0.05263057351112366,
-0.07584847509860992,
0.06380704790353775,
0.02512199431657791,
0.028178859502077103,
-0.006920731160789728,
0.059795111417770386,
-0.0196672473102808,
0.08964395523071289,
0.08038042485713959,
-0.007235884666442871,
0.09868589043617249,
-0.03191833570599556,
0.006547331809997559,
-0.057698819786310196,
0.06932510435581207,
-0.12982366979122162,
0.05436630919575691,
0.043436627835035324,
-0.10945180803537369,
0.03841061517596245,
0.02560393325984478,
0.11603125184774399,
0.058632634580135345,
-0.040632184594869614,
-0.10494323819875717,
-0.13799439370632172,
0.023235952481627464,
0.058803655207157135,
-0.06312531977891922,
-0.13800419867038727,
-0.052970461547374725,
-0.2062724232673645,
0.04198472201824188,
-0.07393307238817215,
0.06842854619026184,
0.045238204300403595,
0.01849091611802578,
-0.05578908324241638,
-0.06200101599097252,
0.01771395653486252,
0.13669656217098236,
-0.06059794872999191,
-0.13932769000530243
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# hubert-rinna-jdrt-RTSPsplit-0206-2
This model is a fine-tuned version of [rinna/japanese-hubert-base](https://huggingface.co/rinna/japanese-hubert-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0758
- Wer: 0.2423
- Cer: 0.0746
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 35
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 14.3788 | 1.0 | 60 | 13.6628 | 0.9855 | 0.9894 |
| 7.2762 | 2.0 | 120 | 6.5080 | 0.9855 | 0.9894 |
| 5.3923 | 3.0 | 180 | 5.1318 | 0.9855 | 0.9894 |
| 4.1787 | 4.0 | 240 | 3.9522 | 0.9855 | 0.9894 |
| 3.3514 | 5.0 | 300 | 3.1901 | 0.9855 | 0.9894 |
| 2.8121 | 6.0 | 360 | 2.6919 | 0.9855 | 0.9894 |
| 2.0411 | 7.0 | 420 | 1.9743 | 1.0 | 0.8450 |
| 1.4368 | 8.0 | 480 | 1.2798 | 1.0 | 0.5637 |
| 1.1834 | 9.0 | 540 | 1.0842 | 1.0 | 0.5095 |
| 1.0832 | 10.0 | 600 | 1.0301 | 0.9903 | 0.5145 |
| 0.9659 | 11.0 | 660 | 0.8236 | 0.8099 | 0.4526 |
| 0.7583 | 12.0 | 720 | 0.7280 | 0.8166 | 0.4551 |
| 0.7525 | 13.0 | 780 | 0.7380 | 0.8118 | 0.4639 |
| 0.7021 | 14.0 | 840 | 0.6548 | 0.8114 | 0.4818 |
| 0.6591 | 15.0 | 900 | 0.6535 | 0.8043 | 0.4555 |
| 0.6141 | 16.0 | 960 | 0.5780 | 0.7767 | 0.3729 |
| 0.5776 | 17.0 | 1020 | 0.5662 | 0.7726 | 0.3717 |
| 1.1193 | 18.0 | 1080 | 0.4865 | 0.7044 | 0.3245 |
| 0.5106 | 19.0 | 1140 | 0.5317 | 0.6884 | 0.2815 |
| 0.4611 | 20.0 | 1200 | 0.4132 | 0.6135 | 0.2433 |
| 0.4474 | 21.0 | 1260 | 0.3655 | 0.5785 | 0.1975 |
| 0.5201 | 22.0 | 1320 | 0.3572 | 0.5594 | 0.1805 |
| 0.3513 | 23.0 | 1380 | 0.2872 | 0.4983 | 0.1652 |
| 0.3245 | 24.0 | 1440 | 0.2417 | 0.4383 | 0.1531 |
| 0.2967 | 25.0 | 1500 | 0.2275 | 0.4089 | 0.1305 |
| 0.2441 | 26.0 | 1560 | 0.1895 | 0.3835 | 0.1419 |
| 0.2236 | 27.0 | 1620 | 0.1519 | 0.3254 | 0.1029 |
| 0.2215 | 28.0 | 1680 | 0.1423 | 0.3254 | 0.1046 |
| 0.1583 | 29.0 | 1740 | 0.1195 | 0.2985 | 0.0955 |
| 0.1627 | 30.0 | 1800 | 0.1101 | 0.2836 | 0.0888 |
| 0.1528 | 31.0 | 1860 | 0.0973 | 0.2680 | 0.0830 |
| 0.1585 | 32.0 | 1920 | 0.0909 | 0.2635 | 0.0828 |
| 0.1443 | 33.0 | 1980 | 0.0825 | 0.2520 | 0.0784 |
| 0.1322 | 34.0 | 2040 | 0.0775 | 0.2467 | 0.0765 |
| 0.1824 | 35.0 | 2100 | 0.0758 | 0.2423 | 0.0746 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "rinna/japanese-hubert-base", "model-index": [{"name": "hubert-rinna-jdrt-RTSPsplit-0206-2", "results": []}]} | automatic-speech-recognition | tndklab/hubert-rinna-jdrt-RTSPsplit-0206-2 | [
"transformers",
"safetensors",
"hubert",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:rinna/japanese-hubert-base",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:37:34+00:00 | [] | [] | TAGS
#transformers #safetensors #hubert #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us
| hubert-rinna-jdrt-RTSPsplit-0206-2
==================================
This model is a fine-tuned version of rinna/japanese-hubert-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0758
* Wer: 0.2423
* Cer: 0.0746
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0002
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 35
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 35",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #hubert #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 35",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
66,
115,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #hubert #automatic-speech-recognition #generated_from_trainer #base_model-rinna/japanese-hubert-base #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 35### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.09041319042444229,
0.10401695966720581,
-0.0023518067318946123,
0.08839446306228638,
0.12425249069929123,
-0.0035006816033273935,
0.13787509500980377,
0.14191928505897522,
-0.06928262114524841,
0.0741477906703949,
0.09589006006717682,
0.10972340404987335,
0.02624472975730896,
0.14655694365501404,
-0.061267588287591934,
-0.2790803909301758,
0.04358288645744324,
0.009770245291292667,
-0.02155330777168274,
0.11185131222009659,
0.09066060930490494,
-0.1251661479473114,
0.08072565495967865,
0.0003617451584432274,
-0.12917013466358185,
0.022951606661081314,
0.018401382490992546,
-0.0982062891125679,
0.12821154296398163,
0.021274132654070854,
0.0823909118771553,
0.027941375970840454,
0.10159891098737717,
-0.23608863353729248,
0.007289950270205736,
0.03480704873800278,
0.033670421689748764,
0.057666752487421036,
0.03650228679180145,
-0.00561425369232893,
0.12231907248497009,
-0.08997638523578644,
0.05340014770627022,
0.04031824320554733,
-0.10989808291196823,
-0.24524827301502228,
-0.06523916125297546,
0.03640547767281532,
0.10084658861160278,
0.09737938642501831,
-0.027848705649375916,
0.09542763233184814,
-0.05875576287508011,
0.08644264936447144,
0.24842624366283417,
-0.32341375946998596,
-0.04952014982700348,
-0.025126321241259575,
0.04850907623767853,
0.07212035357952118,
-0.11026228219270706,
-0.007718770299106836,
0.06399214267730713,
0.018196990713477135,
0.10355959087610245,
-0.036938078701496124,
-0.08755036443471909,
0.0062613110058009624,
-0.12788955867290497,
-0.0037978969048708677,
0.14974063634872437,
0.05279592424631119,
-0.04834664985537529,
-0.09283620119094849,
-0.049587830901145935,
-0.1262483149766922,
-0.06038367375731468,
-0.028026927262544632,
0.05109114572405815,
-0.044431574642658234,
-0.05944793298840523,
-0.02531220018863678,
-0.07283136993646622,
-0.09090316295623779,
-0.030354158952832222,
0.19140122830867767,
0.04878154769539833,
-0.01271010935306549,
-0.0072067128494381905,
0.06857559084892273,
-0.01863659918308258,
-0.14591364562511444,
-0.010447312146425247,
0.03239857405424118,
-0.0037498867604881525,
-0.007908573374152184,
-0.030273083597421646,
0.0032148624304682016,
0.04611288011074066,
0.15574170649051666,
-0.08836556226015091,
0.07466201484203339,
-0.017971016466617584,
0.009769132360816002,
-0.10705740004777908,
0.19615215063095093,
-0.05454491823911667,
-0.058245766907930374,
0.0062963333912193775,
0.09364094585180283,
0.04506698250770569,
-0.009344770573079586,
-0.09112671762704849,
0.007679355330765247,
0.09600132703781128,
0.04414267838001251,
-0.08619150519371033,
0.07241205126047134,
-0.03180911764502525,
0.011381176300346851,
-0.00044140024692751467,
-0.12079513072967529,
0.026974469423294067,
0.024803319945931435,
-0.08019202202558517,
-0.030735481530427933,
0.007692073471844196,
0.011494133621454239,
-0.010849179700016975,
0.07051803916692734,
-0.06793566793203354,
0.018702318891882896,
-0.05406993255019188,
-0.11269285529851913,
0.0065776691772043705,
-0.08675462752580643,
0.026836436241865158,
-0.11478213965892792,
-0.12805283069610596,
-0.014279060997068882,
0.025710197165608406,
-0.04609883576631546,
-0.006760668009519577,
-0.10314983129501343,
-0.09050049632787704,
0.024611303582787514,
-0.02510351501405239,
0.047731149941682816,
-0.08071910589933395,
0.09729945659637451,
0.07584047317504883,
0.08058251440525055,
-0.017407869920134544,
0.03981715813279152,
-0.09775493294000626,
0.017334358766674995,
-0.17761124670505524,
0.054222289472818375,
-0.07894860208034515,
0.035094428807497025,
-0.111271932721138,
-0.08265496045351028,
0.022467004135251045,
0.030484123155474663,
0.0684928223490715,
0.12672488391399384,
-0.15634025633335114,
-0.07833023369312286,
0.1733463704586029,
-0.10295554250478745,
-0.11054801940917969,
0.11748743802309036,
-0.039984289556741714,
0.052849091589450836,
0.06673571467399597,
0.2542373538017273,
0.01560883317142725,
-0.13392485678195953,
-0.001090374425984919,
-0.027323605492711067,
0.06067356467247009,
0.010958967730402946,
0.06032837554812431,
-0.008058467879891396,
-0.0013133910251781344,
0.02988152578473091,
-0.05516950041055679,
0.03359321877360344,
-0.08285709470510483,
-0.08731948584318161,
-0.028008749708533287,
-0.1067148745059967,
0.012607603333890438,
0.03886880725622177,
0.06008518114686012,
-0.1191883534193039,
-0.09456459432840347,
0.027477741241455078,
0.0994124487042427,
-0.10397167503833771,
0.03949404135346413,
-0.10618283599615097,
0.05951518937945366,
-0.005030498839914799,
-0.015110312961041927,
-0.14476272463798523,
0.020920298993587494,
0.03505223989486694,
-0.01715732365846634,
0.046315573155879974,
-0.05577267333865166,
0.07279925048351288,
0.06179798021912575,
-0.06822526454925537,
-0.061219073832035065,
-0.01189235970377922,
0.018385890871286392,
-0.06901519745588303,
-0.20675185322761536,
-0.00857491698116064,
-0.0342305563390255,
0.11760134994983673,
-0.19181609153747559,
0.017283914610743523,
-0.021139245480298996,
0.08722636848688126,
0.036899980157613754,
-0.007724607363343239,
-0.010466395877301693,
0.06926310807466507,
-0.03166952729225159,
-0.0507979691028595,
0.04011363536119461,
-0.004190836101770401,
-0.09417585283517838,
0.016727207228541374,
-0.15769590437412262,
0.13797132670879364,
0.1418319046497345,
-0.03161260113120079,
-0.07707392424345016,
0.010046368464827538,
-0.029843097552657127,
-0.03363756835460663,
-0.028691785410046577,
0.011351954191923141,
0.1567029058933258,
-0.033448025584220886,
0.1414874792098999,
-0.08854787051677704,
0.0004464688536245376,
0.03057747334241867,
-0.04518702253699303,
-0.007889499887824059,
0.12881721556186676,
0.06360132992267609,
-0.06166524067521095,
0.1277306228876114,
0.10733888298273087,
-0.09910966455936432,
0.13960795104503632,
-0.04950384050607681,
-0.07337333261966705,
-0.021149517968297005,
0.01673908159136772,
0.005570319481194019,
0.12141133099794388,
-0.13031227886676788,
-0.02807135321199894,
0.008310571312904358,
0.010834110900759697,
0.012109831906855106,
-0.2043757140636444,
-0.019319213926792145,
0.019873343408107758,
-0.09390100091695786,
-0.00587747897952795,
0.006354092620313168,
-0.008663080632686615,
0.11363232135772705,
-0.013628475368022919,
-0.11174820363521576,
-0.001272851717658341,
-0.010504866018891335,
-0.06830119341611862,
0.17846007645130157,
-0.1082318127155304,
-0.16742919385433197,
-0.10472497344017029,
-0.06600228697061539,
-0.038756512105464935,
0.028338661417365074,
0.07341307401657104,
-0.1143113374710083,
-0.045673880726099014,
-0.1138915941119194,
0.02940882369875908,
0.01471125241369009,
0.027097975835204124,
0.007615221664309502,
0.011851637624204159,
0.06644764542579651,
-0.10910152643918991,
-0.011335194110870361,
-0.04975854977965355,
-0.03505675122141838,
0.022449916228652,
0.029920216649770737,
0.1216968297958374,
0.13678507506847382,
-0.0006002399022690952,
0.019023234024643898,
-0.04273676127195358,
0.18453939259052277,
-0.07767055183649063,
-0.023245850577950478,
0.13028819859027863,
-0.008819776587188244,
0.015037364326417446,
0.16027739644050598,
0.039335381239652634,
-0.11432016640901566,
0.017038773745298386,
0.023665932938456535,
-0.019640499725937843,
-0.21224349737167358,
-0.04124729335308075,
-0.038451600819826126,
0.01821104995906353,
0.08385811001062393,
0.03147789090871811,
0.02544267475605011,
0.02068142034113407,
0.031927961856126785,
0.007774911355227232,
0.02126321941614151,
0.06869830191135406,
0.12735994160175323,
0.03087587095797062,
0.11090581864118576,
-0.03867766633629799,
-0.06147903949022293,
0.025064479559659958,
-0.0027630762197077274,
0.19184407591819763,
0.019633516669273376,
0.1297844797372818,
0.039751868695020676,
0.16006194055080414,
0.011629919521510601,
0.04571576416492462,
0.016839785501360893,
-0.025908004492521286,
-0.014146912842988968,
-0.06089332699775696,
-0.0283554308116436,
0.0620729960501194,
-0.024787046015262604,
0.04924575611948967,
-0.12167565524578094,
0.013223282992839813,
0.05566461384296417,
0.2805137634277344,
0.06372582167387009,
-0.3025754392147064,
-0.07631469517946243,
0.018773280084133148,
-0.06945651769638062,
-0.014757021330296993,
0.07308408617973328,
0.12820647656917572,
-0.06056647375226021,
0.05260089412331581,
-0.037262123078107834,
0.07446902990341187,
-0.04057849571108818,
0.03800356015563011,
0.03588531166315079,
0.07410667836666107,
0.00026363946381025016,
0.03881857916712761,
-0.27824726700782776,
0.29111015796661377,
0.0116238659247756,
0.09002209454774857,
-0.0339190848171711,
-0.0007281063590198755,
0.04191406071186066,
0.019782161340117455,
0.12681534886360168,
-0.04174261912703514,
-0.11631714552640915,
-0.1935945749282837,
-0.05906476452946663,
0.03794817999005318,
0.13420890271663666,
-0.003301377175375819,
0.10922511667013168,
-0.0337277427315712,
-0.024290546774864197,
0.062490515410900116,
-0.07647760212421417,
-0.11142707616090775,
-0.0797874704003334,
-0.035557497292757034,
0.08112742751836777,
0.017000723630189896,
-0.06131460890173912,
-0.0823247879743576,
-0.09245847165584564,
0.10441253334283829,
-0.04072748124599457,
-0.008826046250760555,
-0.09674707800149918,
0.025601103901863098,
0.1186901405453682,
-0.08087871968746185,
0.054090265184640884,
0.02130131423473358,
0.0628916546702385,
0.033837802708148956,
-0.04217992722988129,
0.11438899487257004,
-0.07433914393186569,
-0.1738746464252472,
-0.04866141825914383,
0.1493675410747528,
0.025487961247563362,
0.05477839708328247,
0.0004267183248884976,
0.0189665537327528,
0.012196044437587261,
-0.0635431706905365,
0.02006828784942627,
0.02896573767066002,
0.029847295954823494,
0.03589894622564316,
-0.08045187592506409,
-0.042356401681900024,
-0.10529417544603348,
-0.04056653007864952,
0.14867253601551056,
0.29771333932876587,
-0.0763499066233635,
0.04191388934850693,
0.08240868896245956,
-0.05338317155838013,
-0.19144009053707123,
0.0010301799047738314,
0.028790749609470367,
0.02163834497332573,
0.0024931717198342085,
-0.1508360058069229,
0.0802411362528801,
0.0689508318901062,
-0.025786766782402992,
0.08192933350801468,
-0.2807070016860962,
-0.14536453783512115,
0.1452157348394394,
0.1260337382555008,
0.0880265161395073,
-0.15236781537532806,
-0.03394624590873718,
-0.0202754158526659,
-0.06462167203426361,
0.06226889789104462,
-0.06836643069982529,
0.1345295011997223,
-0.0048312474973499775,
0.040210824459791183,
0.018104378134012222,
-0.037153858691453934,
0.12500308454036713,
0.002367899054661393,
0.08328104764223099,
-0.03979335352778435,
-0.016986338421702385,
-0.006257283501327038,
-0.046122096478939056,
0.06398503482341766,
-0.10057561099529266,
0.03485025092959404,
-0.04283475503325462,
-0.04005751758813858,
-0.07026288658380508,
0.025229042395949364,
-0.009391832165420055,
-0.05962701141834259,
-0.04677773639559746,
0.039928656071424484,
0.04368686303496361,
0.0015281130326911807,
0.1470639556646347,
-0.04031573235988617,
0.1238156408071518,
0.1316363364458084,
0.0981779620051384,
-0.04572542756795883,
0.006983592174947262,
0.01943439431488514,
-0.04105982184410095,
0.07215386629104614,
-0.13265374302864075,
0.04906225949525833,
0.11423086374998093,
0.019959354773163795,
0.15934400260448456,
0.05203013867139816,
-0.044053416699171066,
0.026893731206655502,
0.06395609676837921,
-0.13905394077301025,
-0.11996306478977203,
-0.003720419015735388,
-0.05942394211888313,
-0.05847886949777603,
0.05830518901348114,
0.12551464140415192,
-0.07954064011573792,
-0.003372353967279196,
-0.019066276028752327,
0.02506866306066513,
-0.050865672528743744,
0.19987300038337708,
0.05196690559387207,
0.03389643505215645,
-0.10172101855278015,
0.10091906785964966,
0.03033052571117878,
-0.10638563334941864,
0.053362809121608734,
0.07790911942720413,
-0.08314765244722366,
-0.03604670241475105,
0.02123153582215309,
0.13465085625648499,
0.00023313895508181304,
-0.07775373756885529,
-0.15625286102294922,
-0.12919481098651886,
0.054523956030607224,
0.1829497069120407,
0.07556804269552231,
0.011442150920629501,
-0.035178691148757935,
0.031155161559581757,
-0.11966472119092941,
0.09397066384553909,
0.06219545006752014,
0.052894193679094315,
-0.14259590208530426,
0.11550462245941162,
0.021016094833612442,
0.020520320162177086,
-0.02870412915945053,
0.02394162118434906,
-0.1337602734565735,
0.024834049865603447,
-0.13414564728736877,
0.009509017691016197,
-0.050669897347688675,
0.0007696800748817623,
0.010928324423730373,
-0.0722079947590828,
-0.07909218966960907,
0.021440494805574417,
-0.10058894008398056,
-0.013460054062306881,
0.019436391070485115,
0.06867200881242752,
-0.12433899194002151,
-0.03473346307873726,
0.028137454763054848,
-0.08606819808483124,
0.0782645121216774,
0.07311645895242691,
-0.02217506431043148,
0.08192887902259827,
-0.1346282958984375,
-0.008442453108727932,
0.0872536227107048,
0.003968709148466587,
0.034645307809114456,
-0.12387339025735855,
-0.013835244812071323,
0.010112503543496132,
0.061967529356479645,
0.003997083753347397,
0.09163789451122284,
-0.10817264020442963,
0.003287128871306777,
-0.0645134299993515,
-0.0665736123919487,
-0.056943897157907486,
0.023322194814682007,
0.1222974956035614,
0.005181179381906986,
0.17168104648590088,
-0.11528006196022034,
0.01785379834473133,
-0.17998377978801727,
0.007865579798817635,
-0.028905414044857025,
-0.11206909269094467,
-0.12003961950540543,
-0.043864864856004715,
0.07660825550556183,
-0.06121758371591568,
0.129266619682312,
-0.013337692245841026,
0.04897143319249153,
0.03308921679854393,
-0.10146910697221756,
0.010275047272443771,
0.047660719603300095,
0.23698413372039795,
0.03631098195910454,
-0.03986493870615959,
0.042225271463394165,
0.014259436167776585,
0.10228799283504486,
0.1273147612810135,
0.15020263195037842,
0.2032822221517563,
0.0056180995889008045,
0.13302433490753174,
0.0742894783616066,
-0.06784812361001968,
-0.12376624345779419,
0.08480703085660934,
-0.05476151779294014,
0.08640100061893463,
-0.012008623220026493,
0.2333582639694214,
0.12543237209320068,
-0.1542314738035202,
0.04390490800142288,
-0.041293028742074966,
-0.08474710583686829,
-0.13217276334762573,
-0.03618486970663071,
-0.1150309219956398,
-0.17908839881420135,
0.028380436822772026,
-0.11168230324983597,
0.04757843539118767,
0.05849332734942436,
0.02346317283809185,
-0.0026707034558057785,
0.16608697175979614,
0.005375249776989222,
0.009786388836801052,
0.0907941460609436,
0.01239211205393076,
-0.0554596483707428,
-0.0596342533826828,
-0.09298069775104523,
0.025457343086600304,
-0.02354259416460991,
0.016419164836406708,
-0.02931772544980049,
-0.07937595248222351,
0.05058799684047699,
-0.03428856283426285,
-0.0899416133761406,
0.018898019567131996,
0.022151220589876175,
0.08079476654529572,
0.05889619141817093,
0.04469096288084984,
-0.041161611676216125,
0.013002082705497742,
0.248347669839859,
-0.0968327671289444,
-0.09227974712848663,
-0.11898177862167358,
0.26779040694236755,
0.04253199324011803,
0.024307837709784508,
0.001815110445022583,
-0.07381235808134079,
-0.02217205800116062,
0.2286009043455124,
0.19519755244255066,
-0.06858871132135391,
-0.004026460461318493,
-0.030341312289237976,
0.00023563913418911397,
-0.05669866502285004,
0.07626575231552124,
0.1311458796262741,
0.06042220816016197,
-0.043022993952035904,
-0.0589941032230854,
-0.04656083881855011,
-0.029263416305184364,
-0.03887265548110008,
0.08910917490720749,
0.017854634672403336,
-0.026329239830374718,
-0.04853803291916847,
0.06215681508183479,
-0.08873848617076874,
-0.12985704839229584,
0.026896530762314796,
-0.20179885625839233,
-0.1451909989118576,
-0.0013572281459346414,
0.07678213715553284,
0.037143681198358536,
0.03307247906923294,
-0.02545025572180748,
0.002930345945060253,
0.06438135355710983,
-0.007125052157789469,
-0.06777921319007874,
-0.08449497818946838,
0.05745725706219673,
-0.10749062895774841,
0.21889685094356537,
-0.02317899279296398,
0.04377637803554535,
0.10364113748073578,
0.06722890585660934,
-0.05954176187515259,
0.1278359740972519,
0.056134868413209915,
-0.11425803601741791,
0.0396089069545269,
0.13675916194915771,
-0.04753297194838524,
0.15033945441246033,
0.056655190885066986,
-0.1300361007452011,
0.028582118451595306,
-0.03591402247548103,
-0.09434323757886887,
-0.07358982414007187,
-0.0323609933257103,
-0.06898985058069229,
0.13100562989711761,
0.17891813814640045,
-0.04600270092487335,
0.015039084479212761,
-0.04034971073269844,
0.04791847988963127,
0.05816137045621872,
0.04482417181134224,
-0.03287457302212715,
-0.2597147226333618,
0.002872242359444499,
0.012943532317876816,
-0.012481391429901123,
-0.25071924924850464,
-0.08694405108690262,
-0.008582082577049732,
-0.03960881382226944,
-0.09264318645000458,
0.08407814800739288,
0.1008964478969574,
0.03945621848106384,
-0.04627106338739395,
-0.1258767992258072,
-0.01622917503118515,
0.18643620610237122,
-0.15908503532409668,
-0.09649860113859177
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mi_modelo_small
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.9190
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 156 | 2.0811 |
| No log | 2.0 | 312 | 1.3007 |
| No log | 3.0 | 468 | 0.9190 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "mi_modelo_small", "results": []}]} | question-answering | asier86/mi_modelo_small | [
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"question-answering",
"generated_from_trainer",
"base_model:distilbert-base-uncased",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:37:38+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #distilbert #question-answering #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #endpoints_compatible #region-us
| mi\_modelo\_small
=================
This model is a fine-tuned version of distilbert-base-uncased on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.9190
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #distilbert #question-answering #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
65,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #question-answering #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.10708221793174744,
0.08787639439105988,
-0.0020266782958060503,
0.10252698510885239,
0.1274273842573166,
0.02112007699906826,
0.141371950507164,
0.11134063452482224,
-0.08131598681211472,
0.05829233676195145,
0.13410121202468872,
0.11583946645259857,
-0.0018265031976625323,
0.09120819717645645,
-0.0611470602452755,
-0.2005513310432434,
0.004526883829385042,
0.03799702972173691,
-0.09097661823034286,
0.11530310660600662,
0.08769755810499191,
-0.13717275857925415,
0.07678336650133133,
-0.0056874570436775684,
-0.17407989501953125,
0.026953130960464478,
0.006698115728795528,
-0.037090420722961426,
0.1213178038597107,
0.022234441712498665,
0.13187210261821747,
0.023392172530293465,
0.07922547310590744,
-0.19724278151988983,
0.016064628958702087,
0.06023797392845154,
-0.002444722456857562,
0.07741731405258179,
0.025374948978424072,
0.00522031681612134,
0.08328425884246826,
-0.08128440380096436,
0.06258498877286911,
0.029924798756837845,
-0.12490696460008621,
-0.2442893236875534,
-0.09979312866926193,
0.05106590315699577,
0.09691397845745087,
0.09645190834999084,
-0.012973576784133911,
0.13962721824645996,
-0.07699720561504364,
0.08457877486944199,
0.23785600066184998,
-0.3144238293170929,
-0.07457572966814041,
0.04030787572264671,
0.03424936532974243,
0.07267975062131882,
-0.10186595469713211,
-0.031949128955602646,
0.07634760439395905,
0.02774656005203724,
0.10123031586408615,
-0.03811340406537056,
-0.08034272491931915,
0.02562544122338295,
-0.14547140896320343,
-0.012132234871387482,
0.16400417685508728,
0.06514542549848557,
-0.04620467126369476,
-0.03429178148508072,
-0.06315664947032928,
-0.11850249767303467,
-0.029209623113274574,
-0.038746949285268784,
0.04297671839594841,
-0.036058444529771805,
-0.07818788290023804,
-0.016910377889871597,
-0.1026930958032608,
-0.09884173423051834,
-0.050167523324489594,
0.15796343982219696,
0.04260306805372238,
0.016289860010147095,
-0.024042820557951927,
0.10198324918746948,
-0.033326566219329834,
-0.14423881471157074,
0.002374351257458329,
0.020160887390375137,
-0.005382745061069727,
-0.05033143609762192,
-0.04691127687692642,
-0.05261760577559471,
0.03646983578801155,
0.1889425367116928,
-0.07942666113376617,
0.04144255816936493,
0.027733104303479195,
0.03756984323263168,
-0.09601793438196182,
0.14689302444458008,
-0.0625351294875145,
-0.016724655404686928,
-0.002582760527729988,
0.0813518762588501,
0.03259194642305374,
0.0043426938354969025,
-0.08961918205022812,
0.021114153787493706,
0.09003882110118866,
0.02003106102347374,
-0.04685822129249573,
0.05495336651802063,
-0.05019587278366089,
-0.0007271741633303463,
0.012994267977774143,
-0.08166541904211044,
0.026867516338825226,
0.009509325958788395,
-0.060018569231033325,
-0.040473684668540955,
0.02484549954533577,
0.021615348756313324,
0.020202718675136566,
0.09433086216449738,
-0.0980396494269371,
0.008519732393324375,
-0.08998142182826996,
-0.10547184944152832,
0.022492701187729836,
-0.06569727510213852,
0.033701442182064056,
-0.0928310751914978,
-0.16869968175888062,
-0.013243017718195915,
0.05997863784432411,
-0.03089512512087822,
-0.015934670343995094,
-0.04027499631047249,
-0.09035107493400574,
-0.00003611849751905538,
-0.013057121075689793,
0.09050161391496658,
-0.05944909527897835,
0.10105447471141815,
0.061352185904979706,
0.0693640485405922,
-0.03430663421750069,
0.03159473091363907,
-0.10945126414299011,
0.041360966861248016,
-0.18735328316688538,
0.003874248592182994,
-0.08583753556013107,
0.07179579883813858,
-0.09306249767541885,
-0.08517639338970184,
-0.015431683510541916,
0.0062622493132948875,
0.08747606724500656,
0.10318896174430847,
-0.14944259822368622,
-0.0581885389983654,
0.14982865750789642,
-0.09124893695116043,
-0.17624926567077637,
0.12427528202533722,
-0.047914113849401474,
0.061393290758132935,
0.04755978658795357,
0.16868218779563904,
0.0616682767868042,
-0.11862847954034805,
-0.018172746524214745,
0.003125043585896492,
0.059556711465120316,
-0.027621155604720116,
0.069667287170887,
-0.016716107726097107,
0.003911957144737244,
0.008871469646692276,
-0.06546089053153992,
0.034424856305122375,
-0.09911514818668365,
-0.08977389335632324,
-0.05385984107851982,
-0.10926553606987,
0.04224531352519989,
0.06640573590993881,
0.056745659559965134,
-0.11437124013900757,
-0.10020264238119125,
0.07629279792308807,
0.08758904784917831,
-0.07177864015102386,
0.01953268051147461,
-0.07853355258703232,
0.08141487091779709,
-0.07461990416049957,
-0.0283375047147274,
-0.1568889319896698,
-0.05427051708102226,
0.0030947246123105288,
-0.004757899325340986,
0.006649648770689964,
0.015061174519360065,
0.08364995568990707,
0.06679408252239227,
-0.07277575880289078,
-0.04576537013053894,
-0.05219518020749092,
0.014072518795728683,
-0.1144556924700737,
-0.2103353887796402,
-0.026713289320468903,
-0.03361331671476364,
0.10900475829839706,
-0.21285325288772583,
0.043190985918045044,
-0.0020271677058190107,
0.09231361746788025,
0.036184754222631454,
-0.009037263691425323,
-0.04559910297393799,
0.06368440389633179,
-0.03133980184793472,
-0.06026409938931465,
0.03968898206949234,
0.002780989510938525,
-0.09370192140340805,
-0.07623559981584549,
-0.1157163456082344,
0.1735924631357193,
0.12736397981643677,
-0.09160782396793365,
-0.079501211643219,
-0.0017041253158822656,
-0.06279008835554123,
-0.040141116827726364,
-0.03961801528930664,
0.005880035925656557,
0.11513597518205643,
-0.014275002293288708,
0.11783244460821152,
-0.08347894996404648,
-0.038106340914964676,
0.010217049159109592,
-0.05128463730216026,
0.018388114869594574,
0.1080067977309227,
0.11438195407390594,
-0.07793420553207397,
0.14664483070373535,
0.18690940737724304,
-0.11109071969985962,
0.11495424807071686,
-0.06606268882751465,
-0.08766110986471176,
-0.03343747556209564,
0.023377012461423874,
0.00860913097858429,
0.14631205797195435,
-0.13784223794937134,
0.018673941493034363,
0.015117805451154709,
0.011043009348213673,
0.02062302455306053,
-0.21574048697948456,
-0.04429280757904053,
0.027669047936797142,
-0.04530027136206627,
-0.014141418971121311,
-0.010482629761099815,
-0.008280049078166485,
0.08751733601093292,
-0.012050123885273933,
-0.06216656044125557,
0.039291467517614365,
-0.0028817809652537107,
-0.07470632344484329,
0.2108505219221115,
-0.07383661717176437,
-0.08231393247842789,
-0.11021465808153152,
-0.027316400781273842,
-0.03676888346672058,
0.01191480178385973,
0.06919683516025543,
-0.07928702235221863,
-0.035149816423654556,
-0.09274188429117203,
0.014241885393857956,
0.035347312688827515,
0.028293823823332787,
0.03302577883005142,
0.00197785347700119,
0.09445279836654663,
-0.11119365692138672,
0.007646216079592705,
-0.0434451699256897,
-0.06777635216712952,
0.03244328871369362,
0.04332584887742996,
0.12954339385032654,
0.13176003098487854,
-0.006293661426752806,
-0.003174192737787962,
-0.018926145508885384,
0.24626010656356812,
-0.06957513093948364,
-0.026417139917612076,
0.13200309872627258,
-0.009646299295127392,
0.04485779628157616,
0.1337425708770752,
0.07189981639385223,
-0.0981660783290863,
0.026833930984139442,
0.05833779275417328,
-0.018213531002402306,
-0.22848562896251678,
-0.015088150277733803,
-0.03629239648580551,
0.000045420805690810084,
0.07777763903141022,
0.030681850388646126,
0.040295474231243134,
0.07582857459783554,
0.021042656153440475,
0.045759521424770355,
-0.03376749902963638,
0.0639004036784172,
0.09280970692634583,
0.037512339651584625,
0.11627176403999329,
-0.049132365733385086,
-0.05257520079612732,
0.02913925051689148,
0.003403926733881235,
0.22885066270828247,
0.011365432292222977,
0.13645634055137634,
0.07914537191390991,
0.17988373339176178,
-0.019697509706020355,
0.06641747057437897,
-0.013529207557439804,
-0.064922995865345,
0.0009128026431426406,
-0.05508677661418915,
0.0013120006769895554,
0.04334177449345589,
-0.08823199570178986,
0.08494602143764496,
-0.0997040644288063,
0.025331957265734673,
0.07190584391355515,
0.24650606513023376,
0.05759034305810928,
-0.2938763499259949,
-0.09897506982088089,
0.01800764538347721,
-0.025016997009515762,
-0.01612672209739685,
0.030009277164936066,
0.13152241706848145,
-0.038725271821022034,
0.01120639406144619,
-0.0626465231180191,
0.08149950206279755,
0.011924086138606071,
0.03988318517804146,
0.06834918260574341,
0.08969014137983322,
-0.01107326801866293,
0.07226663827896118,
-0.2856568396091461,
0.271731436252594,
0.01860191859304905,
0.09337102621793747,
-0.04448726028203964,
-0.004649949260056019,
0.018219614401459694,
0.06607591360807419,
0.09256407618522644,
-0.027485761791467667,
-0.04307778924703598,
-0.17113527655601501,
-0.04461004212498665,
0.038364943116903305,
0.09722242504358292,
-0.029817651957273483,
0.10828591138124466,
-0.01972132734954357,
0.01194797083735466,
0.0911933183670044,
-0.012685544788837433,
-0.10899017751216888,
-0.07839783281087875,
-0.016333330422639847,
0.005076844710856676,
-0.04149710386991501,
-0.09716060012578964,
-0.09863841533660889,
-0.1091403141617775,
0.13526444137096405,
-0.04622833430767059,
-0.02194986678659916,
-0.10105779021978378,
0.06971652805805206,
0.09055378288030624,
-0.07615187764167786,
0.03999576345086098,
0.02138366736471653,
0.05533910542726517,
0.03994203358888626,
-0.04782804101705551,
0.11958404630422592,
-0.0817321166396141,
-0.17507250607013702,
-0.061528678983449936,
0.10911136865615845,
0.037768565118312836,
0.044988106936216354,
-0.012711377814412117,
0.004085663706064224,
-0.02373502217233181,
-0.09175744652748108,
0.026362668722867966,
-0.02753734402358532,
0.07150380313396454,
0.0161500945687294,
-0.04852674528956413,
0.04140724986791611,
-0.049215421080589294,
-0.020558204501867294,
0.13631883263587952,
0.29612913727760315,
-0.08491875231266022,
-0.015827137976884842,
0.062439993023872375,
-0.04892180114984512,
-0.19544535875320435,
0.06185584515333176,
0.03257249295711517,
-0.004850971046835184,
0.057877007871866226,
-0.1424984186887741,
0.14537838101387024,
0.10979026556015015,
-0.026831332594156265,
0.11165576428174973,
-0.30763980746269226,
-0.12517096102237701,
0.13092049956321716,
0.15578052401542664,
0.10271019488573074,
-0.17396891117095947,
-0.03480206057429314,
-0.012560687027871609,
-0.1473204791545868,
0.08108969032764435,
-0.1549307405948639,
0.0945407971739769,
-0.013403966091573238,
0.04998783394694328,
0.0037797512486577034,
-0.07027675956487656,
0.14607980847358704,
-0.000046318287786561996,
0.11850225180387497,
-0.04621842876076698,
-0.013441015034914017,
0.07169019430875778,
-0.043754030019044876,
0.03303072974085808,
-0.08953805267810822,
0.05375760793685913,
-0.05097579583525658,
-0.026208076626062393,
-0.06335466355085373,
0.0432647280395031,
-0.05150436982512474,
-0.06759865581989288,
-0.045696575194597244,
0.031161170452833176,
0.02781839855015278,
-0.011523814871907234,
0.14316794276237488,
0.02016613259911537,
0.15032581984996796,
0.11551767587661743,
0.07723074406385422,
-0.06723835319280624,
-0.05477220565080643,
0.011233268305659294,
-0.0318131148815155,
0.07324054837226868,
-0.15704317390918732,
0.041798677295446396,
0.1301533579826355,
0.031892213970422745,
0.13937824964523315,
0.07037341594696045,
-0.04625731706619263,
0.011523851193487644,
0.056804731488227844,
-0.16416457295417786,
-0.13765478134155273,
0.0004902769578620791,
-0.04848967865109444,
-0.14053121209144592,
0.08377081155776978,
0.09766854345798492,
-0.058544453233480453,
0.007872414775192738,
-0.0021101331803947687,
0.004452725872397423,
-0.060293179005384445,
0.1910732388496399,
0.0856165662407875,
0.050420865416526794,
-0.07316560298204422,
0.08543732762336731,
0.025754539296030998,
-0.08191429078578949,
0.0036028979811817408,
0.02972598373889923,
-0.06128942221403122,
-0.04768751561641693,
0.057786114513874054,
0.17645438015460968,
-0.017480891197919846,
-0.05138344317674637,
-0.15388798713684082,
-0.10391058027744293,
0.05379781499505043,
0.17472943663597107,
0.09876082092523575,
0.011043163016438484,
-0.016174687072634697,
0.03864007815718651,
-0.12078005075454712,
0.11567529290914536,
0.039784107357263565,
0.0825420618057251,
-0.15192361176013947,
0.10602233558893204,
-0.005278966389596462,
0.015046751126646996,
-0.021611301228404045,
0.05356408283114433,
-0.12369267642498016,
0.004818740300834179,
-0.16167838871479034,
-0.031912337988615036,
-0.04243628308176994,
-0.00007314696267712861,
0.016739998012781143,
-0.07804747670888901,
-0.0766838788986206,
0.024620801210403442,
-0.10337430983781815,
-0.01428994257003069,
0.06022421643137932,
0.052621982991695404,
-0.14696744084358215,
-0.03189612552523613,
0.041056301444768906,
-0.07057381421327591,
0.0632329210639,
0.0331825315952301,
0.029022322967648506,
0.05176157504320145,
-0.18470416963100433,
0.023166287690401077,
0.05453268811106682,
0.008255518041551113,
0.05301650986075401,
-0.09316183626651764,
-0.029816320165991783,
-0.00498243561014533,
0.055776119232177734,
0.00805442314594984,
0.048523422330617905,
-0.12184276431798935,
-0.006599885877221823,
-0.041501663625240326,
-0.06636733561754227,
-0.06084393337368965,
0.011216291226446629,
0.10473647713661194,
0.0061342609114944935,
0.20113077759742737,
-0.08018026500940323,
0.020927581936120987,
-0.21721096336841583,
0.006357924081385136,
0.0005479813553392887,
-0.07817442715167999,
-0.09612289071083069,
-0.040327563881874084,
0.0530112087726593,
-0.06777244061231613,
0.1314149647951126,
-0.03942599520087242,
0.03305871784687042,
0.036887556314468384,
-0.05675159767270088,
0.06102527305483818,
0.02875526063144207,
0.2587580978870392,
0.017544526606798172,
-0.027191393077373505,
0.012560263276100159,
0.03838387876749039,
0.09434543550014496,
0.07679316401481628,
0.17991912364959717,
0.1845041662454605,
-0.050945721566677094,
0.08982273936271667,
0.054168540984392166,
-0.06128912791609764,
-0.11643441021442413,
0.06968209892511368,
-0.024204589426517487,
0.07028414309024811,
-0.014917990192770958,
0.21106743812561035,
0.11576888710260391,
-0.1583249568939209,
0.013773822225630283,
-0.05545544996857643,
-0.08638123422861099,
-0.10067977756261826,
-0.030203357338905334,
-0.08510877937078476,
-0.17570649087429047,
0.016593754291534424,
-0.1270909309387207,
0.0009135941509157419,
0.09953097254037857,
0.010672074742615223,
-0.014867765828967094,
0.2160998284816742,
0.031276196241378784,
0.053685739636421204,
0.03416353464126587,
-0.0031266885343939066,
-0.03791229426860809,
-0.07161640375852585,
-0.0726158618927002,
0.015160031616687775,
-0.034109894186258316,
0.020006928592920303,
-0.05643845722079277,
-0.05833569914102554,
0.04029514640569687,
-0.005345369689166546,
-0.09742140769958496,
0.0028150181751698256,
0.02954794466495514,
0.04776023328304291,
0.05714746564626694,
0.015563864260911942,
0.023946579545736313,
-0.01110083982348442,
0.21950305998325348,
-0.08350735157728195,
-0.070400170981884,
-0.12050648778676987,
0.1965944468975067,
0.025354061275720596,
0.01319186482578516,
0.017012549564242363,
-0.10124767571687698,
0.030138062313199043,
0.207011878490448,
0.16067950427532196,
-0.08113617449998856,
-0.0012133243726566434,
0.005652322433888912,
-0.009823101572692394,
-0.07486828416585922,
0.05726049095392227,
0.12612278759479523,
0.02443198673427105,
-0.08212122321128845,
-0.06962821632623672,
-0.0440346822142601,
-0.0178349781781435,
-0.041994258761405945,
0.04103654995560646,
0.044749293476343155,
0.007497087586671114,
-0.04636310413479805,
0.06092584878206253,
-0.03342123329639435,
-0.12664338946342468,
0.0767127200961113,
-0.1735985279083252,
-0.14392557740211487,
-0.0183744877576828,
0.12503382563591003,
-0.01228486467152834,
0.053399212658405304,
-0.04180862754583359,
-0.012640140019357204,
0.08096010982990265,
-0.019735140725970268,
-0.06453187018632889,
-0.08844248205423355,
0.08284731209278107,
-0.09417755901813507,
0.23162773251533508,
-0.031115693971514702,
0.0654124766588211,
0.13982054591178894,
0.0345631018280983,
-0.084713414311409,
0.08391895890235901,
0.06642034649848938,
-0.10540223121643066,
0.011755467392504215,
0.07862038165330887,
-0.03128759190440178,
0.120674267411232,
0.05946990102529526,
-0.15106318891048431,
0.0053291283547878265,
-0.03741956129670143,
-0.0729675367474556,
-0.07949789613485336,
-0.031817272305488586,
-0.06557213515043259,
0.13159750401973724,
0.18833664059638977,
-0.04668503627181053,
0.016340022906661034,
-0.04423931986093521,
0.04671303182840347,
0.07153702527284622,
0.053498659282922745,
-0.030362121760845184,
-0.22725465893745422,
0.05558815971016884,
0.05898962914943695,
-0.022746507078409195,
-0.24096186459064484,
-0.09315942227840424,
0.021570716053247452,
-0.05836266651749611,
-0.07721637189388275,
0.06337779760360718,
0.1329357922077179,
0.06411603093147278,
-0.05933794751763344,
-0.11032743752002716,
-0.08574248850345612,
0.15853965282440186,
-0.1339440792798996,
-0.09521511942148209
] |
null | null | null | ERROR: type should be string, got "\nhttps://civitai.com/models/20282/henmixreal\n" | {"license": "mit"} | null | prazzz/henmmix5c | [
"license:mit",
"region:us"
] | 2024-02-06T06:37:50+00:00 | [] | [] | TAGS
#license-mit #region-us
|
URL
| [] | [
"TAGS\n#license-mit #region-us \n"
] | [
11
] | [
"passage: TAGS\n#license-mit #region-us \n"
] | [
0.026221778243780136,
-0.033018264919519424,
-0.008281232789158821,
-0.05295303836464882,
0.052470896393060684,
0.06768012046813965,
0.1598525494337082,
0.04655371606349945,
0.23683255910873413,
-0.05407243221998215,
0.11752297729253769,
0.08923697471618652,
0.004284696187824011,
-0.0009730930323712528,
0.014216204173862934,
-0.17134642601013184,
0.04864625632762909,
-0.02878100797533989,
0.08764812350273132,
0.032233644276857376,
-0.006205103360116482,
-0.03845774009823799,
-0.0022142508532851934,
-0.03178790956735611,
-0.057939812541007996,
0.03869890421628952,
0.045729056000709534,
-0.02754949778318405,
0.14189864695072174,
-0.021783310920000076,
0.13335508108139038,
0.046146418899297714,
-0.011738095432519913,
-0.2486042082309723,
0.008575023151934147,
-0.07252951711416245,
-0.11333522200584412,
0.016201216727495193,
0.035761721432209015,
-0.010069100186228752,
0.032174937427043915,
0.11049123108386993,
-0.011680051684379578,
0.06288356333971024,
-0.2015703022480011,
-0.20486389100551605,
-0.07508610188961029,
-0.07555478066205978,
0.0589042492210865,
0.030872387811541557,
0.05628744140267372,
0.1426718831062317,
-0.18022038042545319,
-0.0018841808196157217,
0.04129622131586075,
-0.3510737717151642,
0.09011197835206985,
0.19666501879692078,
0.06407395005226135,
0.07872317731380463,
-0.04774639382958412,
0.06726468354463577,
0.07745297998189926,
-0.02402484230697155,
-0.10679105669260025,
-0.06142130121588707,
0.040939174592494965,
0.15604156255722046,
-0.03852643445134163,
-0.10356393456459045,
0.2591084837913513,
-0.023262828588485718,
-0.04234466329216957,
0.08201269060373306,
-0.02980397455394268,
-0.040379155427217484,
0.04404358193278313,
0.044016025960445404,
0.036236923187971115,
0.182089164853096,
0.1260262131690979,
-0.03375067934393883,
-0.16269677877426147,
-0.030629513785243034,
-0.2528207004070282,
0.07418664544820786,
-0.003647059667855501,
0.10666298121213913,
-0.20037521421909332,
0.03286786004900932,
-0.15483668446540833,
-0.009493621066212654,
-0.02952384203672409,
-0.059835705906152725,
0.05229754373431206,
-0.0237403754144907,
-0.04600388556718826,
0.07238677144050598,
0.08390641957521439,
0.2046167105436325,
0.023024363443255424,
0.016697337850928307,
-0.10405295342206955,
0.15052515268325806,
0.019140364602208138,
0.024860305711627007,
0.179348424077034,
0.07677878439426422,
-0.04891882464289665,
-0.2251969277858734,
0.027894439175724983,
-0.03671982139348984,
-0.1441805064678192,
0.015881337225437164,
-0.1542915552854538,
0.1736440360546112,
-0.04078168794512749,
-0.06919530034065247,
-0.08578147739171982,
0.09790384024381638,
0.07768166810274124,
-0.021921472623944283,
-0.023105677217245102,
-0.01381723117083311,
0.03522264584898949,
-0.048196230083703995,
-0.11687057465314865,
0.018241960555315018,
0.11869648098945618,
0.12573401629924774,
-0.1483907401561737,
-0.008189842104911804,
-0.017200417816638947,
0.019065292552113533,
0.09696817398071289,
-0.112403005361557,
0.028845038264989853,
-0.09672309458255768,
-0.13033071160316467,
0.036653537303209305,
0.017736904323101044,
-0.019008556380867958,
0.1340927630662918,
0.061849117279052734,
0.056560322642326355,
-0.011025321669876575,
-0.07250872999429703,
-0.14035539329051971,
-0.08679798245429993,
0.1058693379163742,
-0.046787332743406296,
0.010320915840566158,
-0.24556252360343933,
-0.014234079979360104,
-0.14995723962783813,
0.059662189334630966,
-0.0037668521981686354,
-0.08819212019443512,
-0.07740068435668945,
0.21408265829086304,
0.0018596589798107743,
0.04301392287015915,
-0.1078512966632843,
0.054903753101825714,
-0.06764797121286392,
0.10065380483865738,
-0.12895582616329193,
-0.06441528350114822,
0.1613781899213791,
-0.13135331869125366,
-0.14002031087875366,
0.0033312994055449963,
-0.009472889825701714,
0.12053907662630081,
0.0802001804113388,
0.44566696882247925,
-0.058881040662527084,
-0.16201181709766388,
0.1270403116941452,
0.17969723045825958,
-0.13685379922389984,
-0.25928929448127747,
0.12393020838499069,
-0.1636963188648224,
-0.16647985577583313,
0.0040023741312325,
-0.006962866988033056,
0.08049977570772171,
-0.03446655720472336,
-0.056274134665727615,
0.042339932173490524,
0.024350708350539207,
0.029094615951180458,
0.01740112341940403,
0.07037191838026047,
-0.1023021712899208,
0.08444856107234955,
0.058610700070858,
-0.014111426658928394,
0.15077349543571472,
0.011494536884129047,
-0.05393160134553909,
0.014761670492589474,
0.044013332575559616,
-0.015627963468432426,
-0.05899091437458992,
-0.09661509096622467,
0.019826244562864304,
-0.031149597838521004,
0.08229395002126694,
0.1699674129486084,
0.023824702948331833,
-0.02797185815870762,
0.028922779485583305,
0.028606392443180084,
0.1009954959154129,
0.06960704177618027,
0.03099375218153,
-0.04839283227920532,
0.04952205345034599,
-0.0417071171104908,
-0.11430390179157257,
-0.004862460307776928,
-0.011735930107533932,
0.11975742131471634,
-0.08906009048223495,
-0.01223952230066061,
0.05951591953635216,
-0.04513183981180191,
0.0019881438929587603,
0.0428374819457531,
0.0035966038703918457,
0.1388600617647171,
0.004440935328602791,
-0.04352007433772087,
0.17440910637378693,
-0.05288633331656456,
0.15533447265625,
0.1715822070837021,
-0.07049662619829178,
0.015605369582772255,
-0.1273636519908905,
0.003230511210858822,
-0.014480113983154297,
0.05292887985706329,
-0.05400136485695839,
-0.05201306566596031,
-0.01274962443858385,
0.014292534440755844,
-0.03134604170918465,
0.01711403578519821,
-0.06057267636060715,
-0.08167021721601486,
-0.10849859565496445,
0.018649224191904068,
0.20683221518993378,
-0.22544461488723755,
0.1609548032283783,
0.40251004695892334,
0.15190774202346802,
0.21155193448066711,
-0.12478897720575333,
-0.002471078187227249,
-0.06630261242389679,
0.026115071028470993,
-0.024814706295728683,
0.13782677054405212,
-0.13174867630004883,
-0.01413064356893301,
0.03880728408694267,
0.0454997681081295,
0.0661163181066513,
-0.17195898294448853,
-0.15260353684425354,
-0.0034879595041275024,
-0.020591814070940018,
-0.1749730259180069,
0.04874620959162712,
-0.07595308125019073,
0.02181261032819748,
0.018216799944639206,
-0.10832522064447403,
0.16837291419506073,
-0.033566512167453766,
-0.06695768237113953,
0.052613962441682816,
-0.20581911504268646,
-0.07900715619325638,
-0.17772749066352844,
-0.18375012278556824,
0.06050071492791176,
0.05760138854384422,
0.07903145253658295,
-0.05951719731092453,
-0.01922747679054737,
0.061719246208667755,
-0.009363299235701561,
-0.13802112638950348,
-0.04235544428229332,
-0.06993678212165833,
0.08744155615568161,
-0.09474305808544159,
-0.07518411427736282,
-0.07833878695964813,
-0.046996138989925385,
-0.020961694419384003,
0.08125963062047958,
-0.1039251759648323,
0.08903530240058899,
0.1493726521730423,
0.03651920333504677,
0.05440247058868408,
-0.08271230012178421,
0.12693379819393158,
-0.037743739783763885,
-0.09459595382213593,
0.07307634502649307,
0.004350725095719099,
0.04920351505279541,
0.24039287865161896,
0.08962162584066391,
-0.10578162968158722,
-0.01780811697244644,
-0.0968487411737442,
-0.16405464708805084,
-0.2553846538066864,
-0.06823288649320602,
-0.08744750916957855,
0.14417944848537445,
0.014636521227657795,
0.10712126642465591,
0.14313316345214844,
0.01343101728707552,
0.10255914181470871,
-0.08983208239078522,
-0.018939344212412834,
0.031209396198391914,
0.2135104089975357,
-0.05208220332860947,
0.00838248711079359,
-0.13684824109077454,
-0.0256142970174551,
0.14601100981235504,
0.13798639178276062,
0.14503207802772522,
0.31421369314193726,
0.15292863547801971,
0.13410434126853943,
0.13474710285663605,
0.12333164364099503,
0.07403261214494705,
0.03444362059235573,
-0.015304201282560825,
-0.06035377085208893,
-0.003846159903332591,
0.02816268615424633,
0.05421729013323784,
0.06724072247743607,
-0.22906480729579926,
0.041139665991067886,
-0.2661744952201843,
0.03544611483812332,
-0.0854712724685669,
0.1161833181977272,
-0.028890252113342285,
0.11051984131336212,
0.11386284977197647,
0.05553818494081497,
-0.023278791457414627,
0.16036942601203918,
0.032686375081539154,
-0.07703183591365814,
0.020292721688747406,
0.024695809930562973,
0.06633034348487854,
0.08606193959712982,
0.09550496190786362,
-0.020778406411409378,
-0.1831783503293991,
0.025963841006159782,
0.12212833017110825,
-0.20747940242290497,
0.289523184299469,
0.013651901856064796,
-0.0743619054555893,
-0.01690039224922657,
-0.06958060711622238,
0.008433517068624496,
0.12829731404781342,
0.10406835377216339,
0.05508929491043091,
-0.2613787055015564,
-0.13299626111984253,
0.046764206141233444,
-0.00873907096683979,
0.11356569826602936,
-0.0052223424427211285,
-0.14201195538043976,
-0.06640999764204025,
0.05814211815595627,
-0.006591420155018568,
0.13023322820663452,
-0.018290361389517784,
-0.08173255622386932,
-0.010230090469121933,
0.055564697831869125,
-0.001312803477048874,
-0.04580084979534149,
0.07523149996995926,
0.009008137509226799,
0.02259289287030697,
-0.08178020268678665,
0.03887253627181053,
-0.08071476966142654,
-0.25375792384147644,
0.019298138096928596,
-0.04987313598394394,
0.004092312417924404,
-0.04684043675661087,
-0.15448936820030212,
-0.1129264086484909,
-0.15445278584957123,
0.13100723922252655,
-0.03675999864935875,
0.091565802693367,
-0.0817658007144928,
0.13736046850681305,
-0.08521489799022675,
0.05375019088387489,
0.00614814180880785,
0.03918716683983803,
-0.017955513671040535,
-0.1031481996178627,
0.09334362298250198,
-0.1874227225780487,
0.023863423615694046,
0.010427716188132763,
-0.056847453117370605,
-0.01354232057929039,
0.03918023407459259,
-0.08763083070516586,
0.21879427134990692,
0.3331502079963684,
-0.011948764324188232,
0.22546616196632385,
0.35863226652145386,
-0.13763751089572906,
-0.23258967697620392,
-0.1205512136220932,
-0.3263251483440399,
-0.09005610644817352,
0.17321562767028809,
-0.18057219684123993,
0.04850830137729645,
0.16150830686092377,
-0.10868281871080399,
0.22499866783618927,
-0.22723928093910217,
-0.04793389141559601,
0.1823979914188385,
-0.038322996348142624,
0.4527989625930786,
-0.1144307404756546,
-0.1784561723470688,
-0.03637253865599632,
-0.16285361349582672,
0.12426037341356277,
-0.026553882285952568,
0.06700495630502701,
0.02416347898542881,
-0.011372359469532967,
-0.009014161303639412,
-0.04529716446995735,
0.2216065675020218,
0.0522729866206646,
0.10468899458646774,
-0.09159468114376068,
-0.17199653387069702,
0.1907423883676529,
-0.0004908236442133784,
-0.003372655250132084,
-0.05411549657583237,
-0.04850282520055771,
-0.06871756166219711,
0.033092137426137924,
-0.0334564633667469,
0.06195882335305214,
0.03364093229174614,
-0.11903523653745651,
-0.10248823463916779,
0.034111104905605316,
-0.13155671954154968,
-0.054850947111845016,
0.26421889662742615,
-0.02080743946135044,
0.09609334170818329,
0.04959092289209366,
-0.05474294349551201,
-0.13538943231105804,
0.005736751481890678,
-0.07534020394086838,
-0.05711410939693451,
0.06573604047298431,
-0.11453206837177277,
-0.024341827258467674,
0.1293732225894928,
-0.029497180134058,
0.09674722701311111,
0.08061115443706512,
-0.07585363835096359,
0.02032829262316227,
0.15617427229881287,
-0.07247176766395569,
-0.10849180817604065,
0.04999847710132599,
0.04640531167387962,
0.17256882786750793,
0.004101871978491545,
0.02018604800105095,
0.08726977556943893,
0.045959215611219406,
-0.007486662827432156,
0.007311292923986912,
-0.11321697384119034,
-0.04241771996021271,
0.0387241393327713,
-0.005273692775517702,
-0.10946331918239594,
0.16008898615837097,
0.056837860494852066,
0.004653505515307188,
-0.06027700752019882,
0.09720424562692642,
-0.06709636747837067,
-0.07046061009168625,
-0.1753035932779312,
0.018511172384023666,
-0.12734080851078033,
-0.09874535351991653,
0.06846235692501068,
-0.09371624886989594,
-0.04084605351090431,
0.08152704685926437,
0.046927981078624725,
0.14401860535144806,
-0.006597559433430433,
-0.023080874234437943,
0.149825319647789,
-0.0884878933429718,
-0.2241756170988083,
0.01969664730131626,
-0.04083063453435898,
-0.07065816223621368,
-0.0007070365245454013,
0.06069544702768326,
-0.0663156732916832,
-0.11958606541156769,
-0.20477768778800964,
0.10412076860666275,
-0.12043121457099915,
-0.03954985365271568,
-0.1041841059923172,
-0.053260523825883865,
0.07891252636909485,
-0.02613759972155094,
-0.04122013971209526,
-0.047595683485269547,
-0.16630595922470093,
0.054254453629255295,
0.07140932232141495,
0.11125344783067703,
-0.0759999230504036,
-0.018354382365942,
0.1398727148771286,
0.048581548035144806,
0.08479110151529312,
0.07578440010547638,
0.026255371049046516,
0.16728560626506805,
-0.1708206981420517,
-0.0542997270822525,
0.1068294569849968,
-0.026716172695159912,
0.01994573324918747,
0.10631280392408371,
-0.04839588701725006,
0.07042654603719711,
-0.05095988139510155,
0.05859163776040077,
-0.15704534947872162,
-0.13073866069316864,
-0.04184387996792793,
0.023728877305984497,
-0.2260182797908783,
0.015071595087647438,
-0.1769561767578125,
0.19692228734493256,
-0.024228032678365707,
0.11490963399410248,
0.08052190393209457,
0.02052290178835392,
0.03539382666349411,
-0.006019921973347664,
0.00946811307221651,
-0.10524865239858627,
-0.05784677714109421,
-0.07560300827026367,
-0.1168874129652977,
-0.009665017947554588,
0.36614301800727844,
0.02430291846394539,
-0.19682736694812775,
0.051222387701272964,
0.18285293877124786,
0.023639049381017685,
-0.0073763905093073845,
0.26180747151374817,
0.08150359988212585,
-0.023175053298473358,
-0.1782374382019043,
0.0396091528236866,
-0.08699734508991241,
-0.15269799530506134,
0.11385007947683334,
0.09347525984048843,
0.05813581123948097,
0.022930078208446503,
0.10404518246650696,
-0.035940010100603104,
-0.05509711429476738,
-0.13301853835582733,
0.13368983566761017,
-0.001790675800293684,
0.0193882267922163,
0.0897885113954544,
0.19249756634235382,
-0.045275162905454636,
0.05437124893069267,
-0.07336640357971191,
-0.001598604372702539,
-0.15740543603897095,
-0.13358698785305023,
0.06194563955068588,
-0.08269550651311874,
0.06342913210391998,
0.050261519849300385,
0.04341990500688553,
0.31786394119262695,
0.039095040410757065,
-0.046439893543720245,
0.003166865324601531,
-0.14845187962055206,
-0.08075450360774994,
-0.06024569645524025,
-0.03110554814338684,
0.028620192781090736,
-0.13928957283496857,
-0.09898591786623001,
-0.06917677819728851,
-0.130235955119133,
-0.06539803743362427,
0.025270747020840645,
0.014251931570470333,
-0.053083837032318115,
-0.17625881731510162,
-0.04808593541383743,
-0.06644169986248016,
0.10105955600738525,
-0.08462738990783691,
0.1516820639371872,
0.0022449472453445196,
0.030281953513622284,
0.07627002149820328,
0.09585131704807281,
0.018900424242019653,
-0.06975197046995163,
0.05599058046936989,
0.12436293810606003,
0.01323844213038683,
0.1259988248348236,
-0.06034265458583832,
-0.019420607015490532,
-0.014145253226161003,
0.14038437604904175,
0.304447740316391,
-0.01856905221939087,
-0.013814439997076988,
-0.022110093384981155,
0.021388787776231766,
0.10893569141626358,
0.19800719618797302,
-0.03437356278300285,
0.2551359534263611,
-0.058974795043468475,
0.0756678432226181,
-0.013180435635149479,
-0.005362013820558786,
-0.053146667778491974,
0.06074550002813339,
0.06268858164548874,
-0.06877048313617706,
-0.10191375762224197,
0.15178529918193817,
-0.14985080063343048,
0.13306055963039398,
0.14678068459033966,
-0.06057753041386604,
0.03797250986099243,
0.0007459368789568543,
0.19896264374256134,
-0.03570213168859482,
0.0984780564904213,
-0.10653308779001236,
-0.10261140763759613,
-0.14764924347400665,
0.037690844386816025,
-0.36797797679901123,
-0.1756322830915451,
0.11731542646884918,
0.14115898311138153,
0.1759258657693863,
-0.012341637164354324,
0.056479312479496,
0.0033020609989762306,
0.08296097069978714,
-0.04232487455010414,
0.1519634872674942,
0.0612073615193367,
-0.017103128135204315,
-0.15296664834022522,
-0.20328094065189362,
-0.0012039330322295427,
-0.058561209589242935,
0.055583830922842026,
-0.02269243635237217,
0.025347469374537468,
0.07746459543704987,
-0.06768939644098282,
-0.029180381447076797,
-0.02352982573211193,
-0.13262848556041718,
0.052229251712560654,
-0.04354005306959152,
0.0320255309343338,
-0.03958037868142128,
-0.022394726052880287,
-0.039987675845623016,
0.10721533745527267,
-0.22402705252170563,
-0.08517231047153473,
0.1422796994447708,
-0.03421911224722862,
0.1542559564113617,
-0.02848726324737072,
-0.12159585952758789,
-0.024955326691269875,
-0.06977712363004684,
0.10887379199266434,
-0.1419300138950348,
0.038592495024204254,
0.13747453689575195,
0.008710617199540138,
0.031119761988520622,
-0.2533661723136902,
0.050644006580114365,
-0.03556957095861435,
-0.016733208671212196,
-0.057031940668821335
] |
null | null | null |
# Lora of Sugitani Zenjyubou/杉谷善住坊 (Fate/Grand Order)
## What Is This?
This is the LoRA model of waifu Sugitani Zenjyubou/杉谷善住坊 (Fate/Grand Order).
## How Is It Trained?
* This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion).
* The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs).
* The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest).
* Dataset used for training is the `stage3-p480-800` in [CyberHarem/sugitani_zenjubou_fgo](https://huggingface.co/datasets/CyberHarem/sugitani_zenjubou_fgo), which contains 60 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets.
* Trained for 800 steps, 40 checkpoints were saved and evaluated.
* **Trigger word is `sugitani_zenjubou_fgo`.**
* Pruned core tags for this waifu are `brown_hair, breasts, brown_eyes, large_breasts, yellow_eyes, ahoge, ponytail`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
## How to Use It?
### If You Are Using A1111 WebUI v1.7+
**Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 700, you need to download [`700/sugitani_zenjubou_fgo.pt`](https://huggingface.co/CyberHarem/sugitani_zenjubou_fgo/resolve/main/700/sugitani_zenjubou_fgo.pt) as the embedding and [`700/sugitani_zenjubou_fgo.safetensors`](https://huggingface.co/CyberHarem/sugitani_zenjubou_fgo/resolve/main/700/sugitani_zenjubou_fgo.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters.
## Which Step Should I Use?
We selected 5 good steps for you to choose. The best one is step 700.
1480 images (1.52 GiB) were generated for auto-testing.

The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11).
Here are the preview of the recommended steps:
| Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0 | pattern_1 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 |
|-------:|--------:|:----------|:-------------|:--------------|:----------|:---------------------------------------------------------------------------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------|
| 700 | 47 | **0.729** | **0.975** | **0.854** | **0.857** | [Download](https://huggingface.co/CyberHarem/sugitani_zenjubou_fgo/resolve/main/700/sugitani_zenjubou_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 580 | 39 | 0.663 | 0.963 | 0.851 | 0.800 | [Download](https://huggingface.co/CyberHarem/sugitani_zenjubou_fgo/resolve/main/580/sugitani_zenjubou_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 780 | 52 | 0.657 | 0.971 | 0.851 | 0.796 | [Download](https://huggingface.co/CyberHarem/sugitani_zenjubou_fgo/resolve/main/780/sugitani_zenjubou_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 540 | 36 | 0.608 | 0.971 | 0.853 | 0.764 | [Download](https://huggingface.co/CyberHarem/sugitani_zenjubou_fgo/resolve/main/540/sugitani_zenjubou_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 660 | 44 | 0.598 | 0.975 | 0.852 | 0.755 | [Download](https://huggingface.co/CyberHarem/sugitani_zenjubou_fgo/resolve/main/660/sugitani_zenjubou_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
## Anything Else?
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
## All Steps
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* [Steps From 620 to 800](all/0.md)
* [Steps From 420 to 600](all/1.md)
* [Steps From 220 to 400](all/2.md)
* [Steps From 20 to 200](all/3.md)
| {"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/sugitani_zenjubou_fgo"], "pipeline_tag": "text-to-image"} | text-to-image | CyberHarem/sugitani_zenjubou_fgo | [
"art",
"not-for-all-audiences",
"text-to-image",
"dataset:CyberHarem/sugitani_zenjubou_fgo",
"license:mit",
"region:us"
] | 2024-02-06T06:38:49+00:00 | [] | [] | TAGS
#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/sugitani_zenjubou_fgo #license-mit #region-us
| Lora of Sugitani Zenjyubou/杉谷善住坊 (Fate/Grand Order)
===================================================
What Is This?
-------------
This is the LoRA model of waifu Sugitani Zenjyubou/杉谷善住坊 (Fate/Grand Order).
How Is It Trained?
------------------
* This model is trained with HCP-Diffusion.
* The auto-training framework is maintained by DeepGHS Team.
* The base model used for training is deepghs/animefull-latest.
* Dataset used for training is the 'stage3-p480-800' in CyberHarem/sugitani\_zenjubou\_fgo, which contains 60 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets.
* Trained for 800 steps, 40 checkpoints were saved and evaluated.
* Trigger word is 'sugitani\_zenjubou\_fgo'.
* Pruned core tags for this waifu are 'brown\_hair, breasts, brown\_eyes, large\_breasts, yellow\_eyes, ahoge, ponytail'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
How to Use It?
--------------
### If You Are Using A1111 WebUI v1.7+
Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 700, you need to download '700/sugitani\_zenjubou\_fgo.pt' as the embedding and '700/sugitani\_zenjubou\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.
Which Step Should I Use?
------------------------
We selected 5 good steps for you to choose. The best one is step 700.
1480 images (1.52 GiB) were generated for auto-testing.
!Metrics Plot
The base model used for generating preview images is Meina/MeinaMix\_V11.
Here are the preview of the recommended steps:
Anything Else?
--------------
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
All Steps
---------
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* Steps From 620 to 800
* Steps From 420 to 600
* Steps From 220 to 400
* Steps From 20 to 200
| [
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 700, you need to download '700/sugitani\\_zenjubou\\_fgo.pt' as the embedding and '700/sugitani\\_zenjubou\\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 700.\n\n\n1480 images (1.52 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200"
] | [
"TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/sugitani_zenjubou_fgo #license-mit #region-us \n",
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 700, you need to download '700/sugitani\\_zenjubou\\_fgo.pt' as the embedding and '700/sugitani\\_zenjubou\\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 700.\n\n\n1480 images (1.52 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200"
] | [
48,
38,
475
] | [
"passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/sugitani_zenjubou_fgo #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file."
] | [
0.019267594441771507,
-0.0010733518283814192,
-0.0037288328167051077,
0.08688697963953018,
0.09419368952512741,
0.07180163264274597,
0.2193455994129181,
0.07807505130767822,
0.12995243072509766,
-0.061294861137866974,
0.07946423441171646,
0.04244418814778328,
0.0006133655551820993,
0.03854167088866234,
-0.025680910795927048,
-0.18143543601036072,
-0.04944675415754318,
-0.011902345344424248,
-0.007532334420830011,
0.012002155184745789,
0.08209284394979477,
0.009171812795102596,
0.10998967289924622,
-0.047108009457588196,
-0.03768808767199516,
0.04926618933677673,
-0.030304953455924988,
-0.04834027960896492,
0.02747395634651184,
0.0918111801147461,
0.09997352957725525,
0.013519054278731346,
0.05298532918095589,
-0.13812056183815002,
0.0667898952960968,
-0.011706418357789516,
-0.12021128088235855,
-0.0032273398246616125,
0.010307003743946552,
-0.007303211372345686,
0.1300896257162094,
0.03717150539159775,
-0.1255650520324707,
0.04666200652718544,
-0.14800778031349182,
-0.04418416693806648,
-0.04732700064778328,
0.02725072205066681,
0.1293746381998062,
0.07342194765806198,
0.011844114400446415,
0.041960708796978,
-0.06575025618076324,
0.07617496699094772,
0.10820330679416656,
-0.14667898416519165,
-0.0707220807671547,
0.09744581580162048,
0.018840612843632698,
0.13840892910957336,
-0.09830310940742493,
0.09008246660232544,
0.08256765455007553,
-0.05214466527104378,
-0.1637115329504013,
-0.10458525270223618,
-0.18548615276813507,
-0.012929188087582588,
0.009688420221209526,
0.031133681535720825,
0.4252619743347168,
0.07213340699672699,
0.02633165568113327,
0.06660335510969162,
-0.06691474467515945,
0.02562730759382248,
-0.09407198429107666,
0.13445399701595306,
0.037458423525094986,
0.10662215203046799,
-0.0062974742613732815,
-0.10222646594047546,
-0.11281026154756546,
-0.07857885956764221,
-0.08290272206068039,
-0.03646417707204819,
0.013820796273648739,
0.12117516249418259,
-0.19583220779895782,
-0.004671002738177776,
-0.09421859681606293,
-0.13444045186042786,
0.014236644841730595,
-0.09449663013219833,
0.16181078553199768,
0.07249224185943604,
0.007231788709759712,
0.007430131081491709,
0.2568591237068176,
0.11575967818498611,
0.16259364783763885,
0.04959578067064285,
-0.11013151705265045,
0.1295902132987976,
0.04447106644511223,
-0.08508972823619843,
-0.039154231548309326,
-0.08051709830760956,
0.14528751373291016,
-0.057053931057453156,
0.09801115095615387,
-0.05851518735289574,
-0.11667785793542862,
0.0010986877605319023,
-0.11563481390476227,
0.06845999509096146,
0.026366615667939186,
0.01543351635336876,
-0.04384317994117737,
0.04563591629266739,
0.04240656644105911,
-0.03480507433414459,
-0.009792285971343517,
-0.016030659899115562,
-0.0533558651804924,
0.039352502673864365,
0.09547291696071625,
0.030354350805282593,
0.06425535678863525,
-0.009188620373606682,
-0.02666945941746235,
-0.0003768145979847759,
-0.04832543805241585,
0.017246242612600327,
0.05348553508520126,
0.028607983142137527,
0.09201569855213165,
-0.1632087081670761,
-0.0791335180401802,
-0.010314736515283585,
0.03048388659954071,
-0.0015364739811047912,
0.09157351404428482,
-0.00210349028930068,
0.04451781511306763,
-0.007353062275797129,
-0.025270642712712288,
0.05150965601205826,
-0.11155577749013901,
0.09335440397262573,
0.0012244876706972718,
0.09056901186704636,
-0.20787785947322845,
-0.012289718724787235,
-0.04929567500948906,
0.025592703372240067,
0.0547359362244606,
-0.011499167419970036,
-0.09871924668550491,
0.1088874414563179,
-0.0090553667396307,
0.07366298884153366,
-0.11127622425556183,
0.04503004252910614,
0.02161482907831669,
0.09787990897893906,
-0.10244286060333252,
0.010760102421045303,
0.13564889132976532,
-0.16168254613876343,
-0.16067329049110413,
0.10579732060432434,
-0.013757273554801941,
0.04142682999372482,
0.03448110073804855,
0.15613120794296265,
0.15830257534980774,
-0.18689146637916565,
-0.01858355663716793,
0.05208229646086693,
-0.033145733177661896,
-0.06935938447713852,
0.004985288716852665,
0.10556990653276443,
0.013410624116659164,
0.04226314276456833,
-0.03132897615432739,
0.12557877600193024,
-0.03451620414853096,
-0.08022718131542206,
-0.023955218493938446,
-0.0788012444972992,
-0.07935953885316849,
0.04705430567264557,
-0.007486443500965834,
-0.046124719083309174,
0.019456231966614723,
-0.16731421649456024,
0.16404186189174652,
0.01085694320499897,
0.012694158591330051,
-0.08615878224372864,
0.13004809617996216,
0.007995129562914371,
0.007330847904086113,
0.014233357273042202,
-0.05230734497308731,
-0.0939115509390831,
0.24157403409481049,
0.08374616503715515,
0.09150761365890503,
0.06522463262081146,
-0.04141710698604584,
-0.07172566652297974,
0.013482590205967426,
0.007756381761282682,
-0.04513651132583618,
0.03533196076750755,
-0.08718856424093246,
0.06394266337156296,
-0.010973338037729263,
0.025441821664571762,
-0.017360329627990723,
-0.03218400850892067,
0.08703630417585373,
0.021415093913674355,
-0.026287971064448357,
0.09375941008329391,
0.04986384138464928,
-0.0221564918756485,
-0.05333501845598221,
0.011721061542630196,
0.07159700989723206,
-0.008518382906913757,
-0.08435129374265671,
0.036776088178157806,
-0.004111267626285553,
0.08799736201763153,
0.1971013844013214,
-0.19791056215763092,
0.046416524797677994,
0.02391533926129341,
0.040295735001564026,
0.04097071662545204,
-0.0038831436540931463,
-0.029933765530586243,
0.031225787475705147,
-0.03774923458695412,
0.0711723268032074,
-0.02585793286561966,
0.0791928619146347,
-0.01488912757486105,
-0.13778486847877502,
-0.02353687584400177,
-0.008106525987386703,
0.17614443600177765,
-0.18047477304935455,
0.08981698006391525,
0.16364310681819916,
-0.11022519320249557,
0.17440925538539886,
-0.018730126321315765,
-0.013232835568487644,
0.0014925505965948105,
0.044839341193437576,
-0.00033481980790384114,
0.11418423056602478,
-0.1011151522397995,
-0.005852996837347746,
0.023187002167105675,
-0.07774427533149719,
0.03988964110612869,
-0.11351019889116287,
-0.12780873477458954,
-0.07971608638763428,
-0.0434068962931633,
-0.011244887486100197,
0.049755387008190155,
-0.05788692459464073,
0.08069462329149246,
-0.09753801673650742,
-0.06792919337749481,
-0.013773087412118912,
-0.07471391558647156,
0.022931618615984917,
0.0037001646123826504,
-0.0544879212975502,
-0.155158132314682,
-0.10955888032913208,
-0.06081556901335716,
-0.13909882307052612,
-0.0013068291591480374,
0.06226951256394386,
-0.1227698028087616,
-0.04840053990483284,
0.013211253099143505,
-0.06326513737440109,
0.08293570578098297,
-0.07347817718982697,
0.02311706729233265,
0.05625160411000252,
-0.03275429829955101,
-0.15323790907859802,
-0.006613400764763355,
-0.0637863427400589,
-0.07683627307415009,
0.15273655951023102,
-0.1585826426744461,
0.18741440773010254,
-0.04872409626841545,
0.06732525676488876,
0.06650742143392563,
0.03831104934215546,
0.14421318471431732,
-0.12113188952207565,
0.07049493491649628,
0.21674855053424835,
0.05626101419329643,
0.06468065083026886,
0.11487893760204315,
0.06819088011980057,
-0.1164548248052597,
0.03948578983545303,
0.07638067752122879,
-0.10033562034368515,
-0.07280197739601135,
-0.04852890968322754,
-0.11718787252902985,
-0.030374089255928993,
0.04034924507141113,
0.06783145666122437,
0.06529676169157028,
0.11834514141082764,
-0.046433061361312866,
0.008911355398595333,
0.11115379631519318,
0.03662513941526413,
0.06997248530387878,
0.010314330458641052,
0.04831286147236824,
-0.1407572478055954,
-0.05583423003554344,
0.1598043590784073,
0.18079346418380737,
0.1883898377418518,
0.02445375919342041,
0.07985583692789078,
0.11734616011381149,
0.09528782218694687,
0.10684973746538162,
0.03903323784470558,
0.016210271045565605,
0.012828485108911991,
-0.07782764732837677,
-0.04997989907860756,
0.020613063126802444,
-0.003162083914503455,
-0.033862896263599396,
-0.14567717909812927,
0.09684731066226959,
-0.007503203582018614,
0.08755858987569809,
0.10945232957601547,
0.03632507100701332,
-0.09722305834293365,
0.1462765485048294,
0.0872063934803009,
0.09642056375741959,
-0.06960278749465942,
0.13028211891651154,
0.08399372547864914,
-0.019552715122699738,
0.14253567159175873,
0.0349125973880291,
0.14921632409095764,
-0.022205926477909088,
-0.0655909925699234,
-0.07174951583147049,
-0.05281970277428627,
0.004618339240550995,
0.03059433214366436,
-0.24701027572155,
0.0951261818408966,
0.059171076864004135,
0.013884446583688259,
-0.01719989813864231,
-0.057672854512929916,
0.1673804670572281,
0.15515996515750885,
0.10458984225988388,
0.02240757644176483,
-0.03243710845708847,
-0.008131180889904499,
-0.08121995627880096,
0.05013783648610115,
0.020701885223388672,
0.05220888927578926,
-0.05666504427790642,
-0.08866758644580841,
-0.01812942512333393,
-0.0055412547662854195,
0.038297150284051895,
-0.08560670912265778,
-0.1011653020977974,
-0.04642840847373009,
0.25509169697761536,
-0.02182144671678543,
0.04064307361841202,
0.05982217937707901,
0.038228124380111694,
-0.04112027585506439,
0.058421019464731216,
-0.02361365221440792,
-0.029032137244939804,
-0.016806114464998245,
0.01722385548055172,
-0.0023321323096752167,
-0.05429477244615555,
-0.06286688894033432,
-0.03375149145722389,
-0.10543591529130936,
-0.09865280985832214,
0.0070382594130933285,
-0.035764213651418686,
-0.00461013987660408,
-0.020791340619325638,
0.029376551508903503,
-0.08201859891414642,
-0.03623894974589348,
0.02045874111354351,
0.016818266361951828,
-0.04319695383310318,
-0.13721778988838196,
-0.008416036143898964,
-0.019401483237743378,
-0.036824047565460205,
0.029258908703923225,
-0.10265335440635681,
-0.07414137572050095,
-0.05744719132781029,
-0.04410738870501518,
0.10808374732732773,
0.22974927723407745,
-0.014362990856170654,
-0.00983008649200201,
0.14311891794204712,
-0.0834139734506607,
-0.3107442855834961,
-0.18566729128360748,
-0.15465860068798065,
-0.09397642314434052,
0.016197510063648224,
-0.060947805643081665,
0.017160819843411446,
0.09392813593149185,
-0.03895653039216995,
0.23154644668102264,
-0.1807492971420288,
-0.09442916512489319,
0.06861069798469543,
0.09381887316703796,
0.31507840752601624,
-0.25675031542778015,
0.015852952376008034,
-0.10375296324491501,
-0.04314759746193886,
0.03121192753314972,
-0.07381775975227356,
0.1129244938492775,
0.04114730656147003,
0.048558738082647324,
-0.008324867114424706,
-0.007431717123836279,
0.1567680388689041,
-0.0874103307723999,
0.13369300961494446,
-0.12022058665752411,
-0.09617186337709427,
0.20573726296424866,
-0.012944540940225124,
0.0026842912193387747,
-0.18440809845924377,
-0.04687509685754776,
-0.026393026113510132,
0.027898570522665977,
-0.019548289477825165,
0.06240218132734299,
-0.004063529893755913,
-0.022908542305231094,
-0.134705051779747,
0.00899121630936861,
-0.04948234558105469,
0.0653841570019722,
0.2214074432849884,
-0.07466660439968109,
-0.06963551044464111,
0.03326215595006943,
-0.00882491935044527,
0.11701397597789764,
0.019391976296901703,
-0.05523470789194107,
-0.03733045607805252,
0.10041162371635437,
-0.19613757729530334,
0.06351594626903534,
0.00290452572517097,
-0.01257370039820671,
0.03171892836689949,
0.012423518113791943,
-0.003890583524480462,
0.12085457146167755,
0.1738511025905609,
-0.018097758293151855,
-0.04642641916871071,
-0.03804372251033783,
-0.021238667890429497,
0.15083396434783936,
-0.03282906115055084,
0.09984895586967468,
0.012968474999070168,
0.035688240081071854,
0.006819962989538908,
0.058142319321632385,
-0.09769720584154129,
-0.0896064043045044,
0.09282651543617249,
-0.04867083206772804,
-0.08539077639579773,
0.09420342743396759,
0.04511549696326256,
0.07134795933961868,
-0.0009140199981629848,
0.05947340279817581,
0.012927679345011711,
-0.1244489774107933,
-0.000022298760086414404,
0.20310498774051666,
-0.04222891852259636,
-0.0764273852109909,
-0.07496579736471176,
0.011661629192531109,
-0.1206703707575798,
0.08703282475471497,
0.02974620833992958,
-0.020632067695260048,
0.1283223181962967,
-0.04583720117807388,
-0.04432063177227974,
-0.005372885148972273,
-0.054270319640636444,
0.01219780184328556,
-0.14138512313365936,
-0.19817614555358887,
0.05338026583194733,
0.0107668312266469,
-0.062430646270513535,
-0.08721709996461868,
-0.11143993586301804,
0.06262162327766418,
-0.12337451428174973,
0.13848018646240234,
-0.07688277214765549,
0.04853207245469093,
-0.03801878169178963,
-0.057967688888311386,
-0.10352034866809845,
-0.025794999673962593,
-0.048238493502140045,
-0.023919515311717987,
0.06425004452466965,
0.01201262604445219,
-0.11072786897420883,
-0.12063099443912506,
0.06970403343439102,
-0.005622191354632378,
-0.005266466177999973,
0.03357518091797829,
-0.05644330382347107,
0.033897195011377335,
-0.23976430296897888,
-0.042516157031059265,
0.08850329369306564,
0.029265932738780975,
-0.08672045916318893,
0.12625941634178162,
0.047346413135528564,
-0.020086804404854774,
0.03833303600549698,
0.004905357491225004,
0.1478864848613739,
-0.07924271374940872,
0.025587698444724083,
-0.12053048610687256,
-0.1576434075832367,
-0.03493673354387283,
0.029502324759960175,
0.2300606667995453,
0.09181717038154602,
0.10420110821723938,
-0.054774146527051926,
0.03326326981186867,
-0.020337440073490143,
0.06731437891721725,
0.015234125778079033,
-0.09915055334568024,
-0.01817477121949196,
-0.1664356291294098,
-0.06842584908008575,
-0.06547056883573532,
0.15166212618350983,
0.04216120019555092,
-0.12778332829475403,
-0.010271642357110977,
0.08996228128671646,
-0.15849581360816956,
-0.006794587709009647,
0.18309040367603302,
-0.039762839674949646,
0.027730338275432587,
-0.13773249089717865,
0.04189608246088028,
0.07070131599903107,
-0.002046955283731222,
-0.027369637042284012,
0.10495449602603912,
-0.008922118693590164,
0.00325967138633132,
0.023998212069272995,
-0.0360705591738224,
0.100441575050354,
-0.05253475159406662,
0.04331493005156517,
-0.02289591357111931,
-0.04702329263091087,
-0.12353882193565369,
0.1987701803445816,
-0.018619617447257042,
0.015961654484272003,
-0.0433877557516098,
-0.009842499159276485,
-0.09907705336809158,
-0.1386617124080658,
-0.07979969680309296,
-0.12366040796041489,
0.09058304131031036,
-0.06386402994394302,
0.012655594386160374,
-0.012522703036665916,
0.012115474790334702,
-0.08887239545583725,
0.008637555874884129,
-0.15815438330173492,
-0.054790645837783813,
0.026890579611063004,
-0.011848961934447289,
-0.022491762414574623,
-0.0457177497446537,
-0.033431120216846466,
0.017439018934965134,
-0.0677943080663681,
-0.07219913601875305,
0.06024143472313881,
0.0753864273428917,
0.06257551163434982,
-0.15961432456970215,
-0.1024639904499054,
-0.06750413030385971,
0.040869779884815216,
0.07015697658061981,
0.18659989535808563,
0.0346868671476841,
-0.014836899004876614,
0.045923806726932526,
0.13801181316375732,
0.03070356510579586,
-0.06649403274059296,
-0.08101662993431091,
-0.13076476752758026,
-0.13585954904556274,
-0.02128642052412033,
-0.05834396183490753,
-0.021760983392596245,
0.02839355543255806,
0.2542017996311188,
0.18190905451774597,
-0.1505483239889145,
0.041968073695898056,
-0.07437586039304733,
0.03638466075062752,
-0.020754965022206306,
0.14317330718040466,
0.04957234114408493,
0.15729908645153046,
-0.04175707325339317,
-0.017515765503048897,
-0.0727309063076973,
0.01916196011006832,
-0.09448094666004181,
0.03372063487768173,
-0.008948717266321182,
-0.059353236109018326,
-0.058743663132190704,
0.1029505655169487,
-0.13371187448501587,
0.05010497197508812,
0.19978804886341095,
-0.1348024606704712,
-0.0041434005834162235,
-0.043983977288007736,
-0.002639397978782654,
0.11916983872652054,
0.017666669562458992,
-0.08611059933900833,
-0.013731946237385273,
-0.0036196215078234673,
0.03732408210635185,
-0.17908160388469696,
-0.11029411852359772,
-0.013185272924602032,
-0.09558020532131195,
0.1499418169260025,
-0.004507611505687237,
-0.004473157227039337,
0.02861327864229679,
-0.06809452176094055,
-0.0029636668041348457,
0.18668359518051147,
0.015772918239235878,
-0.026533683761954308,
-0.05289985239505768,
-0.06681505590677261,
-0.09395770728588104,
0.0640057772397995,
0.07176334410905838,
0.03623707592487335,
-0.0002502782444935292,
0.16560302674770355,
-0.016115868464112282,
-0.040139954537153244,
0.14938364923000336,
-0.19382692873477936,
0.0927138552069664,
-0.0043974402360618114,
-0.01274438388645649,
-0.07621102780103683,
-0.053054120391607285,
0.06229005381464958,
0.06964413821697235,
-0.17761613428592682,
-0.04509761184453964,
0.07307016849517822,
-0.10011870414018631,
0.08079056441783905,
0.028683673590421677,
-0.07217677682638168,
0.024087656289339066,
-0.1287892758846283,
-0.006473843939602375,
-0.09949880093336105,
0.040743160992860794,
0.17631521821022034,
-0.03818133845925331,
0.017318010330200195,
-0.15943939983844757,
0.05400584638118744,
-0.03372607007622719,
-0.04170828312635422,
-0.07373523712158203
] |
null | null | transformers |
# Model Card for Model ID
* DPO trained on ENERGY-DRINK-LOVE/SOLAR_merge2
## Training Details
### Training Data
* custom collected open-Korean DPO datasets
* remove duplication data
[More Information Needed]
### Training Procedure
DPO | {"license": "apache-2.0", "library_name": "transformers"} | null | ENERGY-DRINK-LOVE/SOLAR_merge2_dpo | [
"transformers",
"safetensors",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:40:32+00:00 | [] | [] | TAGS
#transformers #safetensors #license-apache-2.0 #endpoints_compatible #region-us
|
# Model Card for Model ID
* DPO trained on ENERGY-DRINK-LOVE/SOLAR_merge2
## Training Details
### Training Data
* custom collected open-Korean DPO datasets
* remove duplication data
### Training Procedure
DPO | [
"# Model Card for Model ID\n\n* DPO trained on ENERGY-DRINK-LOVE/SOLAR_merge2",
"## Training Details",
"### Training Data\n\n* custom collected open-Korean DPO datasets\n* remove duplication data",
"### Training Procedure \n\nDPO"
] | [
"TAGS\n#transformers #safetensors #license-apache-2.0 #endpoints_compatible #region-us \n",
"# Model Card for Model ID\n\n* DPO trained on ENERGY-DRINK-LOVE/SOLAR_merge2",
"## Training Details",
"### Training Data\n\n* custom collected open-Korean DPO datasets\n* remove duplication data",
"### Training Procedure \n\nDPO"
] | [
30,
28,
3,
22,
7
] | [
"passage: TAGS\n#transformers #safetensors #license-apache-2.0 #endpoints_compatible #region-us \n# Model Card for Model ID\n\n* DPO trained on ENERGY-DRINK-LOVE/SOLAR_merge2## Training Details### Training Data\n\n* custom collected open-Korean DPO datasets\n* remove duplication data### Training Procedure \n\nDPO"
] | [
-0.07228292524814606,
0.09506982564926147,
-0.0025390717200934887,
0.003201268147677183,
0.2155381441116333,
0.016359560191631317,
0.1747826784849167,
0.016049304977059364,
0.01872035674750805,
-0.048484668135643005,
0.06837460398674011,
0.20537391304969788,
0.1113729402422905,
0.15011128783226013,
-0.03409622237086296,
-0.2572093605995178,
0.13179977238178253,
0.05038962885737419,
-0.07307643443346024,
0.0418546088039875,
0.10650370270013809,
-0.034548066556453705,
0.05954095348715782,
-0.054303936660289764,
-0.060163065791130066,
-0.020713994279503822,
-0.07975795865058899,
-0.07231912761926651,
0.08686965703964233,
-0.0745861753821373,
0.14137586951255798,
0.10242737829685211,
0.018962791189551353,
-0.15557491779327393,
0.0013655135408043861,
0.033506281673908234,
-0.05847727879881859,
0.024364596232771873,
0.019544368609786034,
0.07176893949508667,
0.08283284306526184,
-0.07923038303852081,
0.061928022652864456,
-0.013133666478097439,
0.009946059435606003,
-0.3168759346008301,
-0.10316500067710876,
0.08051696419715881,
0.17975302040576935,
0.08917476236820221,
0.056055162101984024,
0.10579492896795273,
-0.0711904987692833,
-0.006921721622347832,
-0.09613772481679916,
-0.1362358033657074,
0.04029456898570061,
0.23803406953811646,
-0.04415132850408554,
-0.016833549365401268,
0.05021068453788757,
-0.029328245669603348,
0.032133691012859344,
0.026115084066987038,
0.044437069445848465,
-0.01849178597331047,
0.07575621455907822,
0.04800933972001076,
-0.07456675916910172,
0.0023937660735100508,
0.22721423208713531,
0.05886538326740265,
-0.03172384947538376,
-0.11119531095027924,
-0.01971808820962906,
-0.057619545608758926,
-0.022231558337807655,
-0.09375983476638794,
0.0607718899846077,
0.05305572599172592,
0.07844481617212296,
-0.01566084660589695,
-0.0822153389453888,
-0.10659421235322952,
-0.14939475059509277,
0.2109530121088028,
0.05347125604748726,
0.009292513132095337,
-0.09760774672031403,
0.13579541444778442,
-0.007478071842342615,
-0.09068747609853745,
-0.06479189544916153,
-0.10765648633241653,
0.05176980048418045,
-0.04252161085605621,
-0.06665772199630737,
-0.042061030864715576,
0.09376835823059082,
0.11066707223653793,
0.06884726881980896,
-0.07566051185131073,
0.012598596513271332,
0.02456231787800789,
0.018787788227200508,
0.030031325295567513,
-0.11054258793592453,
0.11519403010606766,
0.008461330085992813,
-0.04671778157353401,
0.09022052586078644,
0.03501487895846367,
-0.08855877816677094,
0.029886621981859207,
0.016633447259664536,
0.08595378696918488,
-0.013906148262321949,
0.07437480986118317,
-0.02911907061934471,
-0.015021344646811485,
0.22077834606170654,
-0.04659121111035347,
-0.12273208051919937,
-0.032201509922742844,
-0.03968613222241402,
0.13801705837249756,
0.061768788844347,
0.06837280839681625,
0.055530205368995667,
0.06584081053733826,
-0.006071189418435097,
-0.04739884287118912,
-0.0521991066634655,
-0.021317530423402786,
0.014133556745946407,
-0.07418014109134674,
0.17226621508598328,
-0.16787996888160706,
-0.3433287739753723,
0.04043199121952057,
0.03124443255364895,
-0.02595910243690014,
-0.051345568150281906,
-0.06974582374095917,
-0.041485726833343506,
-0.10966846346855164,
-0.016576755791902542,
0.028530782088637352,
-0.007500581908971071,
0.05725923180580139,
-0.0028196386992931366,
0.017066417261958122,
-0.1456344723701477,
0.07723414152860641,
-0.15590538084506989,
0.03769543021917343,
-0.16002371907234192,
0.004297990817576647,
-0.07981403172016144,
0.0968511551618576,
-0.07233723998069763,
-0.0659731775522232,
-0.021275920793414116,
0.03484455123543739,
-0.022855855524539948,
0.16699707508087158,
-0.12588083744049072,
-0.019398707896471024,
0.10239539295434952,
-0.1523805856704712,
-0.09746991097927094,
-0.007727614138275385,
0.0028923815116286278,
0.11483510583639145,
0.0176254753023386,
0.024265481159090996,
0.05421881377696991,
-0.11718921363353729,
0.030399691313505173,
0.08883895725011826,
-0.14431536197662354,
-0.12842822074890137,
0.1084870845079422,
0.007104001939296722,
-0.10527652502059937,
0.06289511919021606,
-0.12446673959493637,
0.11095630377531052,
-0.04213443025946617,
-0.07856278866529465,
-0.029167095199227333,
-0.14030060172080994,
0.0055802613496780396,
0.017035482451319695,
0.032486557960510254,
0.013245990499854088,
0.013745761476457119,
0.10258325934410095,
0.17808093130588531,
-0.004667846951633692,
-0.08298886567354202,
-0.0628436803817749,
-0.0818902924656868,
-0.09164693206548691,
0.013250486925244331,
0.011896708980202675,
-0.11447098106145859,
-0.007267269305884838,
-0.04012995585799217,
-0.15518809854984283,
0.09810111671686172,
0.06380302459001541,
-0.012585735879838467,
-0.05213845148682594,
-0.04419331252574921,
-0.03860045596957207,
0.1056143119931221,
0.06267378479242325,
-0.08472713083028793,
-0.002138983691111207,
-0.026998765766620636,
0.1499784141778946,
-0.07154399901628494,
0.06993699818849564,
-0.047262806445360184,
0.11566747725009918,
0.020204689353704453,
0.054772909730672836,
0.07699527591466904,
0.0528925321996212,
0.026899265125393867,
-0.07682900875806808,
0.037981875240802765,
0.04349789023399353,
-0.12333564460277557,
0.14878952503204346,
-0.00384069443680346,
0.0643354058265686,
0.2006317526102066,
-0.06035364791750908,
-0.04865172132849693,
0.042879607528448105,
-0.010669896379113197,
-0.04019956663250923,
-0.010299136862158775,
0.0030065460596233606,
-0.10824010521173477,
-0.05585434287786484,
0.09731612354516983,
-0.011562202125787735,
0.013641741126775742,
0.01456782128661871,
0.012114104814827442,
-0.08882920444011688,
0.027753686532378197,
0.16874052584171295,
-0.11487938463687897,
0.22375810146331787,
0.061390891671180725,
0.0040795509703457355,
0.10074785351753235,
-0.07846537977457047,
-0.06803479045629501,
-0.018538029864430428,
0.049386922270059586,
-0.07332894951105118,
0.16261175274848938,
-0.054942864924669266,
0.09753439575433731,
0.11181450635194778,
0.03133667632937431,
0.06889577209949493,
-0.11239386349916458,
-0.13213703036308289,
0.04803278669714928,
-0.028970928862690926,
-0.11283441632986069,
0.1080874502658844,
-0.005559571553021669,
0.05300828441977501,
0.00019976688781753182,
-0.08553333580493927,
0.19727100431919098,
0.03260668367147446,
-0.046537190675735474,
0.14894016087055206,
-0.07470244914293289,
-0.10080387443304062,
-0.04159969463944435,
0.07054229080677032,
-0.008998125791549683,
0.005793166346848011,
0.024402320384979248,
-0.09437188506126404,
-0.05283050984144211,
-0.0021868611220270395,
-0.04797334596514702,
-0.04468199983239174,
-0.005804055370390415,
0.015314565971493721,
0.07375048846006393,
-0.05756141245365143,
-0.10304193943738937,
-0.021977011114358902,
0.03006717935204506,
0.01828838884830475,
0.07175835967063904,
-0.07592512667179108,
0.022585870698094368,
0.03758968785405159,
0.07016174495220184,
-0.015370992012321949,
0.02371899224817753,
-0.006248768884688616,
-0.08292808383703232,
-0.040283963084220886,
0.24218887090682983,
0.07520868629217148,
-0.027235928922891617,
0.08307056874036789,
0.008916746824979782,
-0.04238516464829445,
0.04921985790133476,
-0.009112647734582424,
-0.08256516605615616,
-0.2200571447610855,
-0.08353622257709503,
-0.04790675267577171,
0.09655093401670456,
-0.052190471440553665,
0.0642218291759491,
0.08373959362506866,
0.1393163800239563,
-0.006858890876173973,
0.00958240870386362,
-0.09390565752983093,
0.04278666153550148,
0.0036555284168571234,
0.03272871673107147,
0.002325822366401553,
-0.11556780338287354,
-0.018400337547063828,
0.0560351237654686,
0.11082615703344345,
0.1707753986120224,
0.04446208104491234,
-0.06974279135465622,
0.05013095587491989,
0.1477731615304947,
-0.025914575904607773,
0.12153521180152893,
0.0987405851483345,
-0.04974361136555672,
0.001568721840158105,
-0.016752278432250023,
0.04009828716516495,
0.07288692891597748,
-0.1267436295747757,
-0.058169301599264145,
-0.09996098279953003,
-0.024709532037377357,
-0.00530612375587225,
0.14127139747142792,
-0.043314892798662186,
-0.2351677566766739,
-0.03957364708185196,
0.012350405566394329,
0.025366386398673058,
-0.01414950005710125,
0.010693348944187164,
-0.0047573125921189785,
-0.11623252183198929,
0.14421842992305756,
0.007495407946407795,
-0.004749234300106764,
-0.0389716662466526,
0.002583947964012623,
0.062152616679668427,
-0.008170753717422485,
0.030978938564658165,
0.10536748915910721,
-0.3822522461414337,
0.1434248983860016,
-0.019218752160668373,
0.05195728316903114,
-0.09329278767108917,
0.011045255698263645,
0.008934028446674347,
0.12272738665342331,
0.13207916915416718,
0.003774027107283473,
0.0051217288710176945,
-0.02156861126422882,
-0.051055923104286194,
0.06379938870668411,
-0.028686895966529846,
-0.1030624657869339,
-0.029616642743349075,
0.042919326573610306,
-0.007679218892008066,
0.03744496405124664,
-0.012211187742650509,
-0.12759752571582794,
-0.020480535924434662,
0.03798369690775871,
0.010538761503994465,
0.15419961512088776,
-0.05445953458547592,
-0.007261277642101049,
-0.22597455978393555,
0.11957787722349167,
-0.020126989111304283,
-0.12695889174938202,
-0.08067762851715088,
-0.058332402259111404,
0.00008319462358485907,
0.01819063536822796,
0.03881923854351044,
0.03842887654900551,
-0.03569388389587402,
-0.0020597302354872227,
-0.15026696026325226,
0.10933049768209457,
-0.12729942798614502,
-0.08787006139755249,
0.017648188397288322,
0.01336602121591568,
0.027332544326782227,
-0.049437470734119415,
-0.007568293251097202,
-0.11795624345541,
-0.030903147533535957,
-0.09334531426429749,
0.05691927298903465,
0.04898602142930031,
0.1313055157661438,
0.10230706632137299,
0.0007887117099016905,
-0.02945997565984726,
-0.001179880346171558,
-0.15879125893115997,
0.021244514733552933,
0.16705948114395142,
-0.008120516315102577,
0.02201097272336483,
0.04022867977619171,
0.052079204469919205,
-0.1602441668510437,
-0.05544440075755119,
-0.08289939910173416,
-0.008050364442169666,
0.022861298173666,
-0.23319511115550995,
0.013512485660612583,
0.15347084403038025,
-0.05803526192903519,
0.013005410321056843,
-0.19461946189403534,
-0.07563160359859467,
0.23557230830192566,
0.042941123247146606,
0.3155052363872528,
-0.07879938930273056,
-0.02113664336502552,
-0.03510557860136032,
-0.2818693220615387,
0.22676534950733185,
-0.12382915616035461,
0.015540650114417076,
-0.020016470924019814,
0.053432315587997437,
0.029218729585409164,
-0.02923743613064289,
0.20570233464241028,
0.015978194773197174,
0.11055149883031845,
-0.11691775172948837,
0.031009510159492493,
0.0051537323743104935,
0.01709865964949131,
0.17027691006660461,
-0.04134200140833855,
0.07028588652610779,
-0.09697454422712326,
-0.0738053247332573,
-0.06092927232384682,
0.012419349513947964,
-0.003584605408832431,
-0.15937836468219757,
-0.020257657393813133,
0.06342674791812897,
-0.00294162193313241,
-0.010518299415707588,
0.09027186036109924,
0.06241277977824211,
-0.06847451627254486,
0.007368223741650581,
0.0963144525885582,
0.09453149884939194,
0.10489479452371597,
-0.08663210272789001,
-0.06820215284824371,
0.1024821326136589,
-0.1581381857395172,
0.010917422361671925,
0.021451987326145172,
-0.03389333561062813,
-0.0030883713625371456,
0.04469473659992218,
-0.02620227076113224,
-0.03752569481730461,
0.04447505250573158,
-0.11816463619470596,
-0.22541096806526184,
-0.020900510251522064,
0.0072161126881837845,
0.08093860000371933,
0.2093900889158249,
0.062210481613874435,
-0.10900216549634933,
-0.020024778321385384,
0.005632403772324324,
0.06080976128578186,
-0.134379044175148,
-0.00501071335747838,
0.007673589047044516,
0.04893546923995018,
-0.08196471631526947,
0.16216057538986206,
-0.028579356148838997,
-0.08807577937841415,
0.0012727886205539107,
-0.10681222379207611,
-0.1275179535150528,
-0.09874984622001648,
-0.015613749623298645,
0.1604713499546051,
-0.10898170620203018,
-0.1579427570104599,
-0.089511439204216,
-0.17300765216350555,
0.008502848446369171,
-0.09144026041030884,
0.10360123217105865,
0.11915329098701477,
0.08210350573062897,
-0.034990496933460236,
-0.14405906200408936,
-0.037375785410404205,
-0.08156824111938477,
0.05629628524184227,
-0.15247049927711487,
-0.004022274632006884,
-0.0642211064696312,
0.14323803782463074,
-0.05892314016819,
0.05945153534412384,
-0.03934492543339729,
0.013209340162575245,
-0.089997299015522,
-0.14832206070423126,
-0.07283418625593185,
-0.023789675906300545,
0.033214688301086426,
-0.05945011228322983,
0.01899273507297039,
0.03633003681898117,
-0.06860540062189102,
0.08479273319244385,
0.034089356660842896,
0.033109936863183975,
0.032150380313396454,
-0.043704308569431305,
-0.009938312694430351,
-0.040426772087812424,
0.09045987576246262,
0.1357293725013733,
0.06647896766662598,
0.015540261752903461,
-0.06062695384025574,
0.07919444143772125,
0.057562995702028275,
0.015898732468485832,
0.016074882820248604,
-0.028830312192440033,
-0.04570255056023598,
0.06258462369441986,
-0.15489283204078674,
0.0358470119535923,
0.06970271468162537,
-0.15882670879364014,
-0.07686620950698853,
-0.008947262540459633,
0.0591856986284256,
-0.1019788458943367,
-0.03449723497033119,
0.1207074224948883,
0.023579733446240425,
0.13497520983219147,
-0.01150311715900898,
-0.017162403091788292,
-0.05407511815428734,
-0.018495755270123482,
-0.045154888182878494,
-0.04497439041733742,
0.12558934092521667,
0.05364099517464638,
0.0050230929628014565,
-0.008226349018514156,
0.2685663402080536,
-0.11279577761888504,
-0.1666390597820282,
-0.001352294348180294,
-0.022644637152552605,
0.09832267463207245,
0.03777086362242699,
0.3664519488811493,
0.08198370784521103,
0.018524473533034325,
-0.11722474545240402,
0.06543740630149841,
0.06132784113287926,
-0.05307096615433693,
0.03013877384364605,
0.20230259001255035,
-0.014912839978933334,
0.058765873312950134,
0.0662139430642128,
-0.1302882432937622,
-0.07173368334770203,
-0.011844921857118607,
-0.06327740848064423,
-0.005249740090221167,
0.012624774128198624,
0.06213138997554779,
0.11901658773422241,
-0.06982070952653885,
0.005127300973981619,
0.05465100333094597,
0.009648222476243973,
-0.06178538501262665,
-0.0962410569190979,
-0.01624882034957409,
-0.2080773413181305,
0.03150877356529236,
0.02909419871866703,
-0.10895601660013199,
0.1883256733417511,
-0.0012192116118967533,
-0.10532169044017792,
0.06899727135896683,
0.02291746623814106,
0.027353299781680107,
0.015009554103016853,
-0.03126871958374977,
-0.08084292709827423,
-0.07723535597324371,
-0.17629578709602356,
-0.009277145378291607,
0.03158213943243027,
-0.02751181460916996,
-0.07725383341312408,
-0.021890414878726006,
0.09480713307857513,
-0.002877254970371723,
-0.026257386431097984,
-0.023196807131171227,
0.002214272739365697,
-0.043070368468761444,
0.081894151866436,
-0.027276888489723206,
0.06023634597659111,
0.08629818260669708,
0.1610625684261322,
0.008918240666389465,
-0.08630973845720291,
-0.13262507319450378,
0.11655621230602264,
0.035846687853336334,
0.030371462926268578,
0.10828318446874619,
-0.07029473781585693,
-0.060706041753292084,
0.11443479359149933,
0.16131460666656494,
0.060563553124666214,
0.03610682487487793,
-0.013111214153468609,
-0.020547227934002876,
-0.04141288995742798,
0.012392702512443066,
0.15487761795520782,
-0.012417520396411419,
-0.05076778680086136,
0.025366410613059998,
-0.029739392921328545,
0.024202700704336166,
-0.01785736158490181,
0.10477441549301147,
0.04255203157663345,
-0.1170661672949791,
0.008502276614308357,
0.10922900587320328,
-0.10594677925109863,
-0.01925143413245678,
0.12394881248474121,
0.034862715750932693,
-0.0853973999619484,
-0.09265820682048798,
0.018622420728206635,
-0.0034237403888255358,
-0.014007959514856339,
-0.06860817223787308,
0.05137881264090538,
0.16661806404590607,
0.042951349169015884,
-0.04831937700510025,
-0.05525940656661987,
0.030457163229584694,
0.19898778200149536,
0.06637271493673325,
0.0060663786716759205,
0.06442230194807053,
0.08089955896139145,
0.037178292870521545,
-0.11104784905910492,
0.1673949956893921,
0.01704188995063305,
0.0031028082594275475,
0.08781053870916367,
-0.024406319484114647,
-0.07891201972961426,
-0.03059498220682144,
-0.016557173803448677,
-0.11945110559463501,
-0.03937133774161339,
0.0458327941596508,
-0.09853267669677734,
-0.07822544127702713,
0.09216751903295517,
-0.06024329736828804,
0.05080563575029373,
0.08373595029115677,
-0.06167913228273392,
0.03594934940338135,
-0.056505002081394196,
0.15931330621242523,
0.04146186634898186,
-0.11371966451406479,
-0.03039427287876606,
-0.015878107398748398,
0.018396278843283653,
0.06804170459508896,
0.04363882541656494,
-0.01995624229311943,
0.01582307368516922,
-0.1123659759759903,
-0.09442238509654999,
-0.07943715900182724,
-0.10201657563447952,
0.07662535458803177,
0.01517094299197197,
-0.07755397260189056,
0.03724341094493866,
-0.04548879340291023,
0.05984055995941162,
-0.0932425856590271,
-0.10087579488754272
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GPTuz-finetuned-uzwikitext
*********************************************************************************************************
Bu spaceda text **generation model**ni test(o'rganish maqsadida) fine-tuned qilingan.
Model asosan 50 MB dataset bilan 1.30 minut oralig'ida bajarildi, Agarda modelga fine-tuning qilmoqchi
bo'lsangiz, sizdan kamida 10GB va Google colab Pro version foydalanishni tafsiya qilaman, natia zo'rchiqdi.
*********************************************************************************************************
It achieves the following results on the evaluation set:
- Loss: 2.8346
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 2.4436 | 1.0 | 3206 | 2.9914 |
| 2.2235 | 2.0 | 6412 | 2.8723 |
| 2.1544 | 3.0 | 9618 | 2.8346 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Tokenizers 0.15.1 | {"language": ["uz"], "license": "mit", "pipeline_tag": "text-generation"} | text-generation | ai-nightcoder/GPTuz-finetuned-uzwikitext | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"uz",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:42:28+00:00 | [] | [
"uz"
] | TAGS
#transformers #tensorboard #safetensors #gpt2 #text-generation #uz #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| GPTuz-finetuned-uzwikitext
==========================
*
Bu spaceda text generation modelni test(o'rganish maqsadida) fine-tuned qilingan.
Model asosan 50 MB dataset bilan 1.30 minut oralig'ida bajarildi, Agarda modelga fine-tuning qilmoqchi
bo'lsangiz, sizdan kamida 10GB va Google colab Pro version foydalanishni tafsiya qilaman, natia zo'rchiqdi.
*
It achieves the following results on the evaluation set:
* Loss: 2.8346
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #uz #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Tokenizers 0.15.1"
] | [
59,
98,
4,
27
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #uz #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Tokenizers 0.15.1"
] | [
-0.09045927971601486,
-0.007780480198562145,
-0.0013700980925932527,
0.08111386746168137,
0.16786383092403412,
0.012307386845350266,
0.14631955325603485,
0.102086640894413,
-0.07945239543914795,
0.04729935899376869,
0.13522423803806305,
0.15497030317783356,
0.019745422527194023,
0.1214909702539444,
-0.08240294456481934,
-0.23160430788993835,
0.004439142066985369,
0.02983146160840988,
-0.06485424935817719,
0.11761914193630219,
0.08311466127634048,
-0.13783948123455048,
0.08019036054611206,
-0.012852588668465614,
-0.21798326075077057,
0.004908772185444832,
0.02781430073082447,
-0.06269492954015732,
0.14393360912799835,
0.041936296969652176,
0.12083261460065842,
0.05494115874171257,
0.07736773043870926,
-0.19616267085075378,
0.012690178118646145,
0.03965424373745918,
-0.0038144236896187067,
0.07400747388601303,
0.05615780130028725,
0.015250422060489655,
0.1244349256157875,
-0.10304336249828339,
0.020368501543998718,
0.021541589871048927,
-0.12127494066953659,
-0.1721760332584381,
-0.08033955842256546,
0.009847120381891727,
0.05770178884267807,
0.0867178663611412,
-0.017332762479782104,
0.14102549850940704,
-0.05814088135957718,
0.09333665668964386,
0.22787818312644958,
-0.3269800543785095,
-0.06215923652052879,
0.05690060555934906,
0.05229237303137779,
0.08943061530590057,
-0.1259830743074417,
0.01979617029428482,
0.08504942059516907,
0.024327868595719337,
0.11403258144855499,
-0.03629807010293007,
-0.0325658842921257,
0.009449121542274952,
-0.15911661088466644,
-0.01869427226483822,
0.13483941555023193,
0.027579335495829582,
-0.04608321934938431,
-0.04440128430724144,
-0.05637810379266739,
-0.12765571475028992,
-0.05839700251817703,
-0.02268538810312748,
0.03110211342573166,
-0.03641368821263313,
-0.09238206595182419,
-0.013493401929736137,
-0.12449365854263306,
-0.06940645724534988,
-0.07365026324987411,
0.18980343639850616,
0.04470311105251312,
-0.0007740604924038053,
-0.03465591371059418,
0.0900372639298439,
-0.014679414220154285,
-0.12982960045337677,
0.010414406657218933,
0.03749026730656624,
0.02037634328007698,
-0.056556642055511475,
-0.05841563269495964,
-0.11563040316104889,
0.04330025240778923,
0.09012887626886368,
-0.09508425742387772,
0.05299471691250801,
0.02074596844613552,
0.05555440112948418,
-0.10256559401750565,
0.1939336657524109,
-0.004260190762579441,
-0.003307590028271079,
0.018367361277341843,
0.0587138831615448,
0.02796809747815132,
-0.0157028641551733,
-0.1308836042881012,
-0.0076567381620407104,
0.14008812606334686,
0.026547810062766075,
-0.07255066186189651,
0.09238272160291672,
-0.028675474226474762,
0.0022923042997717857,
-0.0018851273925974965,
-0.08810216188430786,
0.027193089947104454,
0.0017310632392764091,
-0.05229198932647705,
-0.03606089577078819,
0.007573416456580162,
0.007816670462489128,
-0.023764219135046005,
0.1321810930967331,
-0.07364434003829956,
0.031114792451262474,
-0.0813998356461525,
-0.12006282806396484,
0.010027077049016953,
-0.05061829835176468,
0.014296247623860836,
-0.12582966685295105,
-0.13826224207878113,
-0.020089924335479736,
0.027503222227096558,
-0.01614639349281788,
-0.013515572994947433,
-0.06484649330377579,
-0.07259012013673782,
0.01827901415526867,
-0.024043627083301544,
0.12652797996997833,
-0.0564475916326046,
0.12595610320568085,
0.07085791230201721,
0.07058408856391907,
-0.04983333870768547,
0.03153860569000244,
-0.08303936570882797,
0.006669753696769476,
-0.16354039311408997,
0.019606318324804306,
-0.04665796458721161,
0.07654563337564468,
-0.0623244047164917,
-0.08907022327184677,
-0.011413604021072388,
0.021404007449746132,
0.07830152660608292,
0.09384904056787491,
-0.1346288025379181,
-0.077547587454319,
0.1774788647890091,
-0.08351681381464005,
-0.1287015676498413,
0.12860047817230225,
-0.04180959239602089,
0.05426863208413124,
0.08252505958080292,
0.21075157821178436,
0.0372760035097599,
-0.10299651324748993,
0.005579120945185423,
-0.005547141190618277,
0.03531740978360176,
-0.05972028523683548,
0.047655802220106125,
0.012925120070576668,
-0.020711788907647133,
0.030398352071642876,
-0.030026782304048538,
0.060483671724796295,
-0.09145224839448929,
-0.0763644129037857,
-0.04757088050246239,
-0.095624640583992,
0.04701988399028778,
0.06237116456031799,
0.09582222253084183,
-0.12072069197893143,
-0.05803802236914635,
0.053687773644924164,
0.05877700448036194,
-0.07470431923866272,
0.028523139655590057,
-0.06414631009101868,
0.09636066854000092,
-0.051765523850917816,
0.0016240874538198113,
-0.15203091502189636,
-0.022679980844259262,
0.0035864212550222874,
0.02310183458030224,
0.02499014139175415,
0.023530758917331696,
0.07350211590528488,
0.07314170151948929,
-0.05897463113069534,
-0.01777639426290989,
0.010662621818482876,
-0.008161603473126888,
-0.13831442594528198,
-0.1795176863670349,
0.021567024290561676,
-0.02465027943253517,
0.07568517327308655,
-0.2332005500793457,
0.0442449226975441,
0.008565613068640232,
0.04865429922938347,
0.003844347083941102,
0.007984866388142109,
-0.05414166301488876,
0.0751158595085144,
-0.044154394418001175,
-0.03772710636258125,
0.07976573705673218,
0.010789955034852028,
-0.09309801459312439,
-0.01797020435333252,
-0.15603432059288025,
0.20311923325061798,
0.14425918459892273,
-0.11793128401041031,
-0.08553150296211243,
-0.013946490362286568,
-0.04046349227428436,
-0.021674031391739845,
-0.04893668368458748,
0.02517234906554222,
0.1478748470544815,
-0.024866236373782158,
0.1566656231880188,
-0.08093667775392532,
-0.043626535683870316,
0.03651689738035202,
-0.05376069247722626,
0.027244726195931435,
0.11442651599645615,
0.08721105009317398,
-0.12664933502674103,
0.13836446404457092,
0.15881487727165222,
-0.08379516750574112,
0.16689270734786987,
-0.043700478971004486,
-0.0478421151638031,
-0.004198136739432812,
0.03754090517759323,
0.020775845274329185,
0.10016811639070511,
-0.11152902245521545,
-0.011472979560494423,
-0.00028422442846931517,
0.02934419736266136,
0.029413748532533646,
-0.23869745433330536,
-0.04410533234477043,
0.02462330274283886,
-0.05101051926612854,
0.005622814875096083,
-0.0037343138828873634,
-0.018236635252833366,
0.1127646267414093,
0.00007138240471249446,
-0.06049651652574539,
0.03442293405532837,
0.0018549426458775997,
-0.10003678500652313,
0.22636903822422028,
-0.05597361922264099,
-0.17051227390766144,
-0.1131889745593071,
-0.07636537402868271,
-0.0485803447663784,
0.02677195891737938,
0.07517359405755997,
-0.09636000543832779,
-0.030798999592661858,
-0.06714185327291489,
0.02651558630168438,
0.014843958429992199,
0.04641999304294586,
0.0037267571315169334,
0.012489625252783298,
0.058091901242733,
-0.09406770765781403,
-0.020705856382846832,
-0.0682661384344101,
-0.07146666944026947,
0.06277888268232346,
0.028668787330389023,
0.12803390622138977,
0.1456756591796875,
-0.03345116972923279,
0.0044031161814928055,
-0.03646453469991684,
0.20493155717849731,
-0.07490193843841553,
-0.03528489172458649,
0.1327999383211136,
-0.007079227827489376,
0.05361703038215637,
0.11954452842473984,
0.05203181132674217,
-0.10366612672805786,
0.022085007280111313,
0.004762963391840458,
-0.04286918044090271,
-0.208266481757164,
-0.032469455152750015,
-0.05996892973780632,
0.0033362985122948885,
0.07894720137119293,
0.04399775341153145,
0.05057546868920326,
0.0775933787226677,
0.030801944434642792,
0.10377831757068634,
-0.012544285506010056,
0.060028865933418274,
0.11771080642938614,
0.04758274182677269,
0.1340309977531433,
-0.06386972963809967,
-0.0940433219075203,
0.02927592396736145,
-0.030118489637970924,
0.2199149876832962,
0.017761094495654106,
0.11072898656129837,
0.058578722178936005,
0.15876662731170654,
0.030734995380043983,
0.07786020636558533,
-0.015364300459623337,
-0.07882760465145111,
-0.0071124727837741375,
-0.04467034712433815,
-0.024169819429516792,
0.024469897150993347,
-0.07782217115163803,
0.028699925169348717,
-0.12215392291545868,
-0.003422648413106799,
0.06616658717393875,
0.1810552030801773,
0.03576429933309555,
-0.30370137095451355,
-0.079706110060215,
0.015752067789435387,
-0.016561662778258324,
-0.017909657210111618,
0.021384794265031815,
0.13007116317749023,
-0.06674865633249283,
0.04137691110372543,
-0.08150357007980347,
0.08196137845516205,
-0.03588308393955231,
0.06959940493106842,
0.013490932993590832,
0.0970405787229538,
-0.013273476622998714,
0.08453400433063507,
-0.3476163446903229,
0.26548701524734497,
0.0015712880995124578,
0.09542445838451385,
-0.062025658786296844,
-0.018783126026391983,
0.03893997147679329,
0.08998501300811768,
0.05460459738969803,
-0.021360570564866066,
-0.0829990953207016,
-0.18579696118831635,
-0.021584125235676765,
0.04746577516198158,
0.12856365740299225,
-0.00090284418547526,
0.09430694580078125,
-0.04416989907622337,
0.016051145270466805,
0.07910341769456863,
-0.03396432846784592,
-0.0874236524105072,
-0.09038401395082474,
-0.015291078016161919,
0.013644863851368427,
-0.035310011357069016,
-0.0703214779496193,
-0.10186204314231873,
-0.15874983370304108,
0.1717301607131958,
-0.0211617611348629,
-0.015724971890449524,
-0.10209876298904419,
0.08318442106246948,
0.01841104030609131,
-0.09375858306884766,
0.024670550599694252,
0.017940707504749298,
0.06558337807655334,
0.0051910970360040665,
-0.07437773048877716,
0.1146145835518837,
-0.0683857724070549,
-0.1653130054473877,
-0.04885469749569893,
0.08505617082118988,
-0.0002837392967194319,
0.04559336230158806,
0.0014289378887042403,
0.0024222079664468765,
-0.020701156929135323,
-0.10526002943515778,
0.03312849625945091,
-0.026294974610209465,
0.059388261288404465,
-0.011730598285794258,
-0.06136346608400345,
-0.013600490055978298,
-0.07032948732376099,
-0.06722985953092575,
0.2061389535665512,
0.27876219153404236,
-0.07632998377084732,
0.007284534629434347,
0.06379138678312302,
-0.061787400394678116,
-0.25109758973121643,
0.02925199456512928,
0.029377343133091927,
0.01951899565756321,
0.03711703047156334,
-0.12062587589025497,
0.07567164301872253,
0.07761454582214355,
-0.017846301198005676,
0.137312114238739,
-0.3155271112918854,
-0.14735662937164307,
0.10815232992172241,
0.18003889918327332,
0.17955657839775085,
-0.1697498857975006,
-0.04076101630926132,
-0.03770415112376213,
-0.060455322265625,
0.09809205681085587,
-0.09323717653751373,
0.13263054192066193,
-0.02395200915634632,
0.0714220330119133,
0.00442458875477314,
-0.06146136298775673,
0.11508086323738098,
-0.03896097466349602,
0.08850429207086563,
-0.0845302864909172,
-0.02775835245847702,
0.016915667802095413,
-0.04751013591885567,
0.003976855892688036,
-0.12255729734897614,
0.025182418525218964,
-0.017184197902679443,
-0.03722687065601349,
-0.058957889676094055,
0.06111722066998482,
-0.03225165605545044,
-0.056376244872808456,
-0.05357705056667328,
0.006867594551295042,
0.024606313556432724,
-0.014629371464252472,
0.1473492830991745,
-0.002060449682176113,
0.1465839296579361,
0.12110880017280579,
0.08248589932918549,
-0.06769124418497086,
0.004597298800945282,
-0.009800620377063751,
-0.026392841711640358,
0.05229177325963974,
-0.15636642277240753,
0.024008244276046753,
0.1273132562637329,
0.010375035926699638,
0.1410583257675171,
0.08971142023801804,
-0.03769112750887871,
0.022334646433591843,
0.08644184470176697,
-0.16276870667934418,
-0.12513183057308197,
-0.022185150533914566,
-0.05643601343035698,
-0.07586247473955154,
0.0754484161734581,
0.09173084795475006,
-0.07807275652885437,
0.012644774280488491,
-0.022551458328962326,
-0.0022535580210387707,
-0.044773347675800323,
0.2024676650762558,
0.07105899602174759,
0.04758140817284584,
-0.0770839974284172,
0.053161438554525375,
0.030639568343758583,
-0.050635166466236115,
0.01675751619040966,
0.0750720277428627,
-0.09230805188417435,
-0.053959786891937256,
0.09325675666332245,
0.20036707818508148,
-0.05205466225743294,
-0.056779779493808746,
-0.15109588205814362,
-0.13729359209537506,
0.055979639291763306,
0.17478206753730774,
0.0994119867682457,
0.005025320220738649,
-0.0350329615175724,
0.01969166100025177,
-0.1379873901605606,
0.0903240516781807,
0.02437768317759037,
0.08544228225946426,
-0.14518938958644867,
0.14994266629219055,
0.011667844839394093,
0.047238435596227646,
-0.039544228464365005,
0.04261023923754692,
-0.13960596919059753,
0.009871517308056355,
-0.13225242495536804,
-0.012913982383906841,
-0.01936120167374611,
-0.007339916191995144,
-0.0025008937809616327,
-0.06517273187637329,
-0.06447511166334152,
0.010517816059291363,
-0.11391601711511612,
-0.01161283440887928,
0.02547268196940422,
0.0465741865336895,
-0.10405220091342926,
-0.034140583127737045,
0.018023397773504257,
-0.061734821647405624,
0.05018707737326622,
0.02405460923910141,
0.018936239182949066,
0.09325122088193893,
-0.19796092808246613,
0.01495255809277296,
0.0776100903749466,
0.009603144600987434,
0.047440771013498306,
-0.059074562042951584,
-0.0042106639593839645,
0.016543613746762276,
0.0802711546421051,
0.03994017094373703,
0.07380503416061401,
-0.13591541349887848,
0.0226746778935194,
-0.07227741181850433,
-0.06953409314155579,
-0.06326847523450851,
0.040226809680461884,
0.037029046565294266,
0.00820879079401493,
0.18754130601882935,
-0.11672022938728333,
0.015038945712149143,
-0.2042977213859558,
0.0142583092674613,
0.0020880582742393017,
-0.13596390187740326,
-0.10988084971904755,
-0.07081548869609833,
0.06652948260307312,
-0.05054353550076485,
0.12172695994377136,
0.01683943159878254,
0.027062993496656418,
0.030914559960365295,
-0.02743123285472393,
0.020797504112124443,
0.0158173106610775,
0.2443625032901764,
0.033365052193403244,
-0.029997313395142555,
0.0008754928130656481,
0.06088025122880936,
0.11270356178283691,
0.06458447873592377,
0.2037842571735382,
0.13984407484531403,
-0.07846318185329437,
0.1120305135846138,
0.015814263373613358,
-0.03798164054751396,
-0.12307386100292206,
0.06948623806238174,
-0.08380857110023499,
0.06960389763116837,
-0.019916551187634468,
0.19337640702724457,
0.13214123249053955,
-0.15357369184494019,
0.015766330063343048,
-0.06067635864019394,
-0.0904511883854866,
-0.10969384759664536,
-0.06001056730747223,
-0.10380207747220993,
-0.1439010351896286,
0.020057640969753265,
-0.10040462017059326,
0.018273497000336647,
0.0829739198088646,
0.016165977343916893,
-0.026204684749245644,
0.18361444771289825,
0.037733402103185654,
0.0168329905718565,
0.0635402724146843,
-0.021340491250157356,
-0.03605332225561142,
-0.07654377818107605,
-0.07177747786045074,
-0.008922204375267029,
-0.026631982997059822,
0.045912597328424454,
-0.03791310638189316,
-0.057469192892313004,
0.044962622225284576,
-0.025880804285407066,
-0.09492848068475723,
0.009175974875688553,
0.04010751470923424,
0.08264543116092682,
0.03390858322381973,
0.01767267845571041,
-0.005018778145313263,
-0.0014886169228702784,
0.22108767926692963,
-0.06256803870201111,
-0.11904798448085785,
-0.06180759146809578,
0.24725669622421265,
0.05253882706165314,
0.0202290341258049,
0.006019740831106901,
-0.07779774069786072,
0.026547575369477272,
0.2910791039466858,
0.1935153603553772,
-0.09365998208522797,
0.007634143345057964,
-0.015143882483243942,
-0.0010036738822236657,
-0.012554863467812538,
0.13204605877399445,
0.11508439481258392,
0.07606462389230728,
-0.08861816674470901,
-0.04132967069745064,
-0.04872715473175049,
0.0022167579736560583,
-0.04515519365668297,
0.07486181706190109,
0.05406836420297623,
0.01446269080042839,
-0.05412041395902634,
0.06399162113666534,
-0.06535017490386963,
-0.08272580057382584,
0.029100829735398293,
-0.1891874372959137,
-0.13357563316822052,
-0.0013450036058202386,
0.1485157161951065,
-0.030745789408683777,
0.05964772030711174,
-0.04193747788667679,
0.015628356486558914,
-0.017253680154681206,
-0.027224937453866005,
-0.06893660873174667,
-0.011789267882704735,
0.04772191122174263,
-0.09459288418292999,
0.17459730803966522,
-0.045592766255140305,
0.07238724827766418,
0.12944768369197845,
0.050439268350601196,
-0.053069837391376495,
0.10795599222183228,
0.0229684729129076,
-0.083933986723423,
0.05413701385259628,
0.09233663231134415,
-0.030723022297024727,
0.028250087052583694,
0.06111365929245949,
-0.12789151072502136,
0.03643704205751419,
-0.07561756670475006,
-0.047453623265028,
-0.03882720693945885,
-0.03001345507800579,
-0.05909458547830582,
0.1262608915567398,
0.17783223092556,
-0.012164524756371975,
0.027207577601075172,
-0.06864802539348602,
0.025628546252846718,
0.04457637295126915,
0.0592060424387455,
-0.04396223649382591,
-0.2538081407546997,
0.005332751665264368,
0.09524539113044739,
-0.023388704285025597,
-0.29613202810287476,
-0.07783418893814087,
-0.020447487011551857,
-0.04851650819182396,
-0.11549104750156403,
0.09342215955257416,
0.13505041599273682,
0.047914665192365646,
-0.0611206591129303,
-0.13420920073986053,
-0.07529454678297043,
0.15719833970069885,
-0.1229647770524025,
-0.10275384783744812
] |
null | null | transformers | # 🗿 ruGPT-3.5 13B
Language model for Russian. Model has 13B parameters as you can guess from it's name. This is our biggest model so far and it was used for trainig GigaChat (read more about it in the [article](https://habr.com/ru/companies/sberbank/articles/730108/)).
## Dataset
Model was pretrained on a 300Gb of various domains, than additionaly trained on the 100 Gb of code and legal documets. Here is the dataset structure:

Training data was deduplicated, the text deduplication includes 64-bit hashing of each text in the corpus for keeping texts with a unique hash. We also filter the documents based on their text compression rate using zlib4. The most strongly and weakly compressing deduplicated texts are discarded.
## Technical details
Model was trained using Deepspeed and Megatron libraries, on 300B tokens dataset for 3 epochs, around 45 days on 512 V100. After that model was finetuned 1 epoch with sequence length 2048 around 20 days on 200 GPU A100 on additional data (see above).
After the final training perplexity for this model was around 8.8 for Russian.

## Examples of usage
Try different generation strategies to reach better results.
```python
request = "Стих про программиста может быть таким:"
encoded_input = tokenizer(request, return_tensors='pt', \
add_special_tokens=False).to('cuda:0')
output = model.generate(
**encoded_input,
num_beams=2,
do_sample=True,
max_new_tokens=100
)
print(tokenizer.decode(output[0], skip_special_tokens=True))
```
```
>>> Стих про программиста может быть таким:
Программист сидит в кресле,
Стих сочиняет он про любовь,
Он пишет, пишет, пишет, пишет...
И не выходит ни черта!
```
```python
request = "Нейронная сеть — это"
encoded_input = tokenizer(request, return_tensors='pt', \
add_special_tokens=False).to('cuda:0')
output = model.generate(
**encoded_input,
num_beams=4,
do_sample=True,
max_new_tokens=100
)
print(tokenizer.decode(output[0], skip_special_tokens=True))
```
```
>>> Нейронная сеть — это математическая модель, состоящая из большого
количества нейронов, соединенных между собой электрическими связями.
Нейронная сеть может быть смоделирована на компьютере, и с ее помощью
можно решать задачи, которые не поддаются решению с помощью традиционных
математических методов.
```
```python
request = "Гагарин полетел в космос в"
encoded_input = tokenizer(request, return_tensors='pt', \
add_special_tokens=False).to('cuda:0')
output = model.generate(
**encoded_input,
num_beams=2,
do_sample=True,
max_new_tokens=100
)
print(tokenizer.decode(output[0], skip_special_tokens=True))
```
```
>>> Гагарин полетел в космос в 1961 году. Это было первое в истории
человечества космическое путешествие. Юрий Гагарин совершил его
на космическом корабле Восток-1. Корабль был запущен с космодрома
Байконур.
``` | {"language": ["en", "ru"], "license": "mit", "tags": ["gpt3", "transformers"]} | text-generation | KrafterDen/copy | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"gpt3",
"en",
"ru",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:44:09+00:00 | [] | [
"en",
"ru"
] | TAGS
#transformers #pytorch #gpt2 #text-generation #gpt3 #en #ru #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # ruGPT-3.5 13B
Language model for Russian. Model has 13B parameters as you can guess from it's name. This is our biggest model so far and it was used for trainig GigaChat (read more about it in the article).
## Dataset
Model was pretrained on a 300Gb of various domains, than additionaly trained on the 100 Gb of code and legal documets. Here is the dataset structure:
.
After the final training perplexity for this model was around 8.8 for Russian.
.",
"## Dataset\n\nModel was pretrained on a 300Gb of various domains, than additionaly trained on the 100 Gb of code and legal documets. Here is the dataset structure:\n\n.\n\nAfter the final training perplexity for this model was around 8.8 for Russian.\n\n.",
"## Dataset\n\nModel was pretrained on a 300Gb of various domains, than additionaly trained on the 100 Gb of code and legal documets. Here is the dataset structure:\n\n.\n\nAfter the final training perplexity for this model was around 8.8 for Russian.\n\n.## Dataset\n\nModel was pretrained on a 300Gb of various domains, than additionaly trained on the 100 Gb of code and legal documets. Here is the dataset structure:\n\n.\n\nAfter the final training perplexity for this model was around 8.8 for Russian.\n\n (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | tyson0420/stack-llama-2-web | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:45:24+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06061961501836777,
0.15481999516487122,
-0.004844071343541145,
0.02074851468205452,
0.0983177199959755,
0.007407687604427338,
0.07119518518447876,
0.11185134947299957,
-0.023851769044995308,
0.1167980208992958,
0.031993988901376724,
0.09781743586063385,
0.11217817664146423,
0.16186554729938507,
0.0015333457849919796,
-0.22897611558437347,
0.049678247421979904,
-0.125278040766716,
-0.0294334813952446,
0.11977242678403854,
0.1422213912010193,
-0.10954539477825165,
0.0752737894654274,
-0.038042325526475906,
-0.005828251596540213,
-0.0323176346719265,
-0.06205610930919647,
-0.05266609415411949,
0.05311284959316254,
0.06794639676809311,
0.07308239489793777,
0.01171939354389906,
0.09106900542974472,
-0.2724283039569855,
0.02348201349377632,
0.0805930644273758,
-0.0006441773730330169,
0.07586129754781723,
0.04993962123990059,
-0.08749990910291672,
0.07524524629116058,
-0.060156844556331635,
0.1498761922121048,
0.07955671846866608,
-0.09018243104219437,
-0.19217631220817566,
-0.07921334356069565,
0.09916994720697403,
0.1890910118818283,
0.05953684076666832,
-0.026427440345287323,
0.11642678081989288,
-0.08593545109033585,
0.013638701289892197,
0.06446459144353867,
-0.06054406240582466,
-0.055855002254247665,
0.06904532760381699,
0.08335285633802414,
0.08567540347576141,
-0.12976622581481934,
-0.010767064057290554,
0.015032444149255753,
0.008952446281909943,
0.08948688954114914,
0.017146794125437737,
0.1335189938545227,
0.040557652711868286,
-0.13501930236816406,
-0.043155476450920105,
0.09761431813240051,
0.03665134683251381,
-0.04888195917010307,
-0.2485782504081726,
-0.023432478308677673,
-0.04339504987001419,
-0.03198111802339554,
-0.03649339824914932,
0.043764639645814896,
-0.014506848528981209,
0.07738617807626724,
-0.004502781666815281,
-0.0837155357003212,
-0.04301247000694275,
0.07241875678300858,
0.06128999963402748,
0.02571401372551918,
-0.015821760520339012,
0.0059297760017216206,
0.12327717989683151,
0.11431120336055756,
-0.126715749502182,
-0.052547648549079895,
-0.06306339055299759,
-0.08449548482894897,
-0.044861067086458206,
0.030838407576084137,
0.037995077669620514,
0.045936476439237595,
0.23867325484752655,
0.007765117567032576,
0.053257301449775696,
0.04455438256263733,
0.014407169073820114,
0.06501194834709167,
0.11008983850479126,
-0.05894824117422104,
-0.09719445556402206,
-0.028582042083144188,
0.10156717151403427,
0.007986726239323616,
-0.04139331728219986,
-0.05712985619902611,
0.07059531658887863,
0.018587570637464523,
0.12360043078660965,
0.08000938594341278,
0.003056557849049568,
-0.0755772516131401,
-0.062465377151966095,
0.17764076590538025,
-0.15825673937797546,
0.04532013460993767,
0.03055616281926632,
-0.0341108962893486,
-0.009745313785970211,
0.012105142697691917,
0.025474950671195984,
-0.021481726318597794,
0.09522198140621185,
-0.05601342022418976,
-0.034448131918907166,
-0.11389608681201935,
-0.03694311901926994,
0.030394554138183594,
0.011153047904372215,
-0.02865210548043251,
-0.03502652049064636,
-0.08865131437778473,
-0.06405586749315262,
0.09101516753435135,
-0.07148737460374832,
-0.04784895107150078,
-0.016645915806293488,
-0.07833752781152725,
0.021804187446832657,
0.01691517047584057,
0.09064167737960815,
-0.0222476739436388,
0.03985358029603958,
-0.0550384595990181,
0.061440225690603256,
0.11723454296588898,
0.027987057343125343,
-0.05787884071469307,
0.061519939452409744,
-0.2424532175064087,
0.10252492874860764,
-0.07715212553739548,
0.04971238598227501,
-0.15203025937080383,
-0.02478341944515705,
0.03986154496669769,
0.01284773275256157,
-0.008251311257481575,
0.14196595549583435,
-0.21994100511074066,
-0.030957341194152832,
0.16964265704154968,
-0.10025953501462936,
-0.08109250664710999,
0.060782887041568756,
-0.05354252830147743,
0.11210215091705322,
0.04557164013385773,
-0.02375967986881733,
0.05775221437215805,
-0.14725260436534882,
-0.011030761525034904,
-0.041942402720451355,
-0.0180682260543108,
0.16207332909107208,
0.0703711211681366,
-0.06047816202044487,
0.07456906884908676,
0.01960151270031929,
-0.014246034435927868,
-0.04887177795171738,
-0.02822130173444748,
-0.1047162413597107,
0.01184528972953558,
-0.06102835759520531,
0.018109694123268127,
-0.021768750622868538,
-0.09445013850927353,
-0.029118487611413002,
-0.17402999103069305,
-0.0031633328180760145,
0.08821269869804382,
-0.011630427092313766,
-0.021509924903512,
-0.11245372891426086,
0.009332616813480854,
0.030967719852924347,
0.0002618339203763753,
-0.13677829504013062,
-0.06033218279480934,
0.026970699429512024,
-0.16097871959209442,
0.029791243374347687,
-0.05741601809859276,
0.04530094936490059,
0.04005871340632439,
-0.03433511033654213,
-0.03489551320672035,
0.010874404571950436,
0.010431389324367046,
-0.01894843392074108,
-0.25422003865242004,
-0.01882786676287651,
-0.0234990194439888,
0.1751047968864441,
-0.22956320643424988,
0.042598169296979904,
0.07489731162786484,
0.1460893303155899,
0.007349682506173849,
-0.03550100699067116,
0.015185600146651268,
-0.07262228429317474,
-0.03268764168024063,
-0.06316669285297394,
-0.01207790058106184,
-0.038400664925575256,
-0.05820201337337494,
0.04906858503818512,
-0.1686294972896576,
-0.030321966856718063,
0.10717973858118057,
0.06342670321464539,
-0.1473218947649002,
-0.02780107781291008,
-0.04056945815682411,
-0.04624456167221069,
-0.06676914542913437,
-0.05461418256163597,
0.11812574416399002,
0.056411582976579666,
0.04860803112387657,
-0.07140495628118515,
-0.07455260306596756,
0.008036690764129162,
-0.01956399530172348,
-0.014917809516191483,
0.09334591031074524,
0.07554110884666443,
-0.12264352291822433,
0.09177418053150177,
0.09668384492397308,
0.08576478064060211,
0.10314212739467621,
-0.014663571491837502,
-0.08914592862129211,
-0.040637146681547165,
0.02245822176337242,
0.016187267377972603,
0.15129362046718597,
-0.012961224652826786,
0.055492039769887924,
0.0358695350587368,
-0.014034898020327091,
0.011105312965810299,
-0.09736533463001251,
0.02655916102230549,
0.030835967510938644,
-0.016302183270454407,
0.03745110332965851,
-0.0447014644742012,
0.019208140671253204,
0.09039704501628876,
0.040895868092775345,
0.040978945791721344,
0.010155045427381992,
-0.04354988783597946,
-0.11037563532590866,
0.1787576973438263,
-0.12389461696147919,
-0.24818050861358643,
-0.13812170922756195,
0.010281167924404144,
0.04737642779946327,
-0.010411068797111511,
0.006690691225230694,
-0.06616118550300598,
-0.1175973042845726,
-0.09878289699554443,
0.018617089837789536,
0.045352302491664886,
-0.07590975612401962,
-0.06842505931854248,
0.06414616107940674,
0.03875524550676346,
-0.13939815759658813,
0.024007495492696762,
0.04662325978279114,
-0.08205481618642807,
-0.0029386086389422417,
0.0791812464594841,
0.06965780258178711,
0.17661017179489136,
0.013885351829230785,
-0.023669935762882233,
0.026634456589818,
0.20819635689258575,
-0.1436755359172821,
0.10975687950849533,
0.13545554876327515,
-0.08767466992139816,
0.08120133727788925,
0.1998777538537979,
0.03777998685836792,
-0.10680917650461197,
0.03608465939760208,
0.028374753892421722,
-0.028325283899903297,
-0.2502254545688629,
-0.06958996504545212,
0.0019060121849179268,
-0.05172049254179001,
0.07064855098724365,
0.08791537582874298,
0.09593888372182846,
0.016860228031873703,
-0.09976044297218323,
-0.07697858661413193,
0.046900223940610886,
0.10824491083621979,
-0.00015424020239152014,
-0.015208319760859013,
0.0904119610786438,
-0.03033481352031231,
0.01743943803012371,
0.09215071052312851,
0.0030607767403125763,
0.17535938322544098,
0.051709048449993134,
0.17189906537532806,
0.07866133749485016,
0.06444311141967773,
0.02004685252904892,
0.007725914940237999,
0.021817529574036598,
0.017227526754140854,
-0.0030957073904573917,
-0.08709781616926193,
-0.0034981227945536375,
0.1202581599354744,
0.049845851957798004,
0.029173865914344788,
0.012042860500514507,
-0.030704669654369354,
0.08337877690792084,
0.1770893782377243,
0.0029054484330117702,
-0.1893385946750641,
-0.07169844210147858,
0.07795937359333038,
-0.08648337423801422,
-0.10729733109474182,
-0.029470939189195633,
0.041069481521844864,
-0.1729043871164322,
0.016882894560694695,
-0.019335895776748657,
0.10788324475288391,
-0.13190391659736633,
-0.01772487722337246,
0.05657728388905525,
0.06932812184095383,
-0.009677323512732983,
0.06694949418306351,
-0.16090403497219086,
0.11770165711641312,
0.01751571334898472,
0.06636732816696167,
-0.09608277678489685,
0.09618937969207764,
-0.007830657996237278,
0.0041499207727611065,
0.1410749852657318,
0.010120149701833725,
-0.05952107161283493,
-0.09608154743909836,
-0.10546442121267319,
-0.009841260500252247,
0.1306990385055542,
-0.14852415025234222,
0.08813067525625229,
-0.02661319263279438,
-0.044553373008966446,
0.003614129964262247,
-0.12497276812791824,
-0.13103094696998596,
-0.18366187810897827,
0.05707118660211563,
-0.12947207689285278,
0.04045100137591362,
-0.10902881622314453,
-0.045833900570869446,
-0.02098964899778366,
0.20040063560009003,
-0.23137451708316803,
-0.06714103370904922,
-0.1551055610179901,
-0.08061286807060242,
0.14446212351322174,
-0.046455029398202896,
0.08550118654966354,
0.0008278203313238919,
0.19068008661270142,
0.021319707855582237,
-0.017237508669495583,
0.1072206199169159,
-0.10052918642759323,
-0.2010865956544876,
-0.09273224323987961,
0.15895552933216095,
0.13766798377037048,
0.03809428587555885,
-0.004381525795906782,
0.03171157464385033,
-0.02098114788532257,
-0.12076930701732635,
0.020226983353495598,
0.17317426204681396,
0.08982043713331223,
0.025265544652938843,
-0.02972041629254818,
-0.11267432570457458,
-0.07061342149972916,
-0.03774050623178482,
0.024755435064435005,
0.18072067201137543,
-0.07222156971693039,
0.18405316770076752,
0.13775517046451569,
-0.05534014105796814,
-0.19904261827468872,
0.021996473893523216,
0.04293542355298996,
0.0070380112156271935,
0.0323902890086174,
-0.20307663083076477,
0.09384101629257202,
0.0008334947633557022,
-0.05131231248378754,
0.1379684954881668,
-0.1823476254940033,
-0.151598259806633,
0.06042521819472313,
0.043563615530729294,
-0.19374065101146698,
-0.12374074012041092,
-0.08848230540752411,
-0.04693066328763962,
-0.15487661957740784,
0.10312657803297043,
0.0020827590487897396,
0.008401188999414444,
0.03778626397252083,
0.02252252586185932,
0.012139533646404743,
-0.04198719933629036,
0.1914343535900116,
-0.025891713798046112,
0.03347287327051163,
-0.0790715217590332,
-0.060851071029901505,
0.062408581376075745,
-0.058187782764434814,
0.0755455270409584,
-0.025226406753063202,
0.015947066247463226,
-0.10598332434892654,
-0.048235729336738586,
-0.02852320298552513,
0.019321219995617867,
-0.09431382268667221,
-0.09348297864198685,
-0.04829427972435951,
0.09367614984512329,
0.09042316675186157,
-0.03652578964829445,
-0.03649144619703293,
-0.078715980052948,
0.038977332413196564,
0.17627815902233124,
0.18159319460391998,
0.04659178853034973,
-0.07959239184856415,
-0.001915142871439457,
-0.014336181804537773,
0.04684065282344818,
-0.22077152132987976,
0.060553863644599915,
0.04557652771472931,
0.016117896884679794,
0.11537692695856094,
-0.0208132341504097,
-0.16198977828025818,
-0.06710557639598846,
0.061360616236925125,
-0.06944561004638672,
-0.17825035750865936,
0.0039279889315366745,
0.07344977557659149,
-0.16578389704227448,
-0.037031736224889755,
0.04200848564505577,
-0.01189455483108759,
-0.0403641052544117,
0.012352054007351398,
0.08063354343175888,
0.007078902795910835,
0.07699975371360779,
0.055281639099121094,
0.09124495089054108,
-0.10227900743484497,
0.07410510629415512,
0.08149529248476028,
-0.08644098788499832,
0.030720343813300133,
0.09573426842689514,
-0.06469762325286865,
-0.0346054881811142,
0.04237886518239975,
0.08354541659355164,
0.024281201884150505,
-0.04682289808988571,
0.0023111123591661453,
-0.09734189510345459,
0.05927345156669617,
0.11483542621135712,
0.03496333956718445,
0.011234734207391739,
0.03813567012548447,
0.04486291855573654,
-0.08093374222517014,
0.11926916986703873,
0.023795632645487785,
0.020354853942990303,
-0.04112942889332771,
-0.040553025901317596,
0.035851649940013885,
-0.026020776480436325,
-0.011440055444836617,
-0.035174157470464706,
-0.0722682997584343,
-0.014069457538425922,
-0.16000694036483765,
-0.0076758842915296555,
-0.03660871088504791,
0.005114538595080376,
0.022510098293423653,
-0.03652830421924591,
0.00792311318218708,
0.012217256240546703,
-0.06868947297334671,
-0.05553458258509636,
-0.023233558982610703,
0.09422210603952408,
-0.16494666039943695,
0.0220257006585598,
0.0823851153254509,
-0.12121747434139252,
0.09289738535881042,
0.016782134771347046,
0.00412249518558383,
0.026962365955114365,
-0.1545863002538681,
0.04763968288898468,
-0.020152103155851364,
0.013473534025251865,
0.04222847521305084,
-0.21637047827243805,
-0.004404853098094463,
-0.04015503451228142,
-0.05566934496164322,
-0.008993052877485752,
-0.0319182425737381,
-0.11338426172733307,
0.09645436704158783,
0.011025024577975273,
-0.08443772792816162,
-0.02965564839541912,
0.03353232145309448,
0.07690354436635971,
-0.027447547763586044,
0.1498211771249771,
-0.004663881380110979,
0.07559948414564133,
-0.17581342160701752,
-0.02282017655670643,
-0.011197620071470737,
0.022367527708411217,
-0.021871577948331833,
-0.01622559316456318,
0.04623444378376007,
-0.02704801969230175,
0.19120801985263824,
-0.024701936170458794,
0.049393873661756516,
0.06364397704601288,
0.009232889860868454,
-0.013832193799316883,
0.11151392012834549,
0.05708572641015053,
0.024334950372576714,
0.022262847051024437,
0.003451440716162324,
-0.04008655622601509,
-0.009981024079024792,
-0.18596695363521576,
0.06803664565086365,
0.14585918188095093,
0.09060460329055786,
-0.012669353745877743,
0.0707244873046875,
-0.10161512345075607,
-0.12005364894866943,
0.10127941519021988,
-0.06415384262800217,
-0.010188822634518147,
-0.06542414426803589,
0.14027701318264008,
0.14953285455703735,
-0.1886233240365982,
0.06583356112241745,
-0.06602055579423904,
-0.0566304549574852,
-0.11457879096269608,
-0.1930263340473175,
-0.057075321674346924,
-0.050602465867996216,
-0.018466074019670486,
-0.05384097993373871,
0.06939727067947388,
0.05750798434019089,
0.01126816775649786,
0.00868057832121849,
0.08568526059389114,
-0.009656033478677273,
0.00248199631460011,
0.030120067298412323,
0.06713981181383133,
0.016768986359238625,
-0.0321255661547184,
0.0179112758487463,
-0.00597198773175478,
0.034156378358602524,
0.059282708913087845,
0.03608176112174988,
-0.028436895459890366,
0.015559280291199684,
-0.034912437200546265,
-0.11309733241796494,
0.042801856994628906,
-0.029640642926096916,
-0.0749855786561966,
0.1347348988056183,
0.026981467381119728,
0.005015076603740454,
-0.023140020668506622,
0.2503887414932251,
-0.07436972856521606,
-0.09334370493888855,
-0.14373961091041565,
0.11701542884111404,
-0.04212593287229538,
0.0635172426700592,
0.03596310690045357,
-0.10810714215040207,
0.017985546961426735,
0.1320217251777649,
0.15442703664302826,
-0.04732590913772583,
0.019251897931098938,
0.028577854856848717,
0.00439635943621397,
-0.04075566306710243,
0.05177190154790878,
0.07100846618413925,
0.14500564336776733,
-0.05157303810119629,
0.08530787378549576,
0.002609728369861841,
-0.1021018698811531,
-0.041973695158958435,
0.11415864527225494,
-0.014296893030405045,
0.017620453611016273,
-0.057136841118335724,
0.124222531914711,
-0.05874236673116684,
-0.23697422444820404,
0.06316976249217987,
-0.0765061303973198,
-0.1432730257511139,
-0.024886758998036385,
0.071670763194561,
-0.016632623970508575,
0.02605951391160488,
0.07167234271764755,
-0.0754380151629448,
0.18880942463874817,
0.03957989811897278,
-0.05233397334814072,
-0.05954399332404137,
0.0744764655828476,
-0.11850855499505997,
0.27879106998443604,
0.010482731275260448,
0.051307905465364456,
0.1042102724313736,
-0.02021743729710579,
-0.13270841538906097,
0.023401619866490364,
0.09579801559448242,
-0.08917027711868286,
0.04087764397263527,
0.21448291838169098,
-0.00629545608535409,
0.11935057491064072,
0.07611140608787537,
-0.07468950748443604,
0.047562725841999054,
-0.11468592286109924,
-0.07639975845813751,
-0.08699081838130951,
0.09244474768638611,
-0.06785612553358078,
0.14258281886577606,
0.12599852681159973,
-0.05530165135860443,
0.011584274470806122,
-0.028389399871230125,
0.045467376708984375,
0.005578654818236828,
0.100032277405262,
0.011115525849163532,
-0.18496567010879517,
0.024811718612909317,
0.016259413212537766,
0.10884406417608261,
-0.18112654983997345,
-0.09105053544044495,
0.046958595514297485,
0.0005061255069449544,
-0.06443515419960022,
0.12483241409063339,
0.057313691824674606,
0.04654949903488159,
-0.0451689288020134,
-0.026830285787582397,
-0.006042256020009518,
0.14264579117298126,
-0.10707559436559677,
-0.005129707511514425
] |
null | null | null | Original model: https://huggingface.co/DAMO-NLP-MT/polylm-chat-13b
# Model Card for PolyLM-Multialpaca
This model is finetuned on [polyLM-13b](https://huggingface.co/DAMO-NLP-MT/polylm-13b) using the following datasets:
# Demo
[Open](https://modelscope.cn/studios/damo/demo-polylm-multialpaca-13b/summary)
# Bias, Risks, and Limitations
The information below in this section are copied from the model's [official model card](https://arxiv.org/pdf/2307.06018.pdf):
> Our contributions are fully methodological: adding the support of multilingualism to LLM during training and SFT phases. It is unavoidable that PolyLM might exhibit several common deficiencies of language models, e.g. hallucination and toxicity. PolyLM should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application.
# Citation
**BibTeX:**
```bibtex
@misc{wei2023polylm,
title={PolyLM: An Open Source Polyglot Large Language Model},
author={Xiangpeng Wei and Haoran Wei and Huan Lin and Tianhao Li and Pei Zhang and Xingzhang Ren and Mei Li and Yu Wan and Zhiwei Cao and Binbin Xie and Tianxiang Hu and Shangjie Li and Binyuan Hui and Bowen Yu and Dayiheng Liu and Baosong Yang and Fei Huang and Jun Xie},
year={2023},
eprint={2307.06018},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
| {"license": "apache-2.0"} | null | Sosnitskij/polylm-chat-13b-gguf | [
"gguf",
"arxiv:2307.06018",
"license:apache-2.0",
"region:us"
] | 2024-02-06T06:45:50+00:00 | [
"2307.06018"
] | [] | TAGS
#gguf #arxiv-2307.06018 #license-apache-2.0 #region-us
| Original model: URL
# Model Card for PolyLM-Multialpaca
This model is finetuned on polyLM-13b using the following datasets:
# Demo
Open
# Bias, Risks, and Limitations
The information below in this section are copied from the model's official model card:
> Our contributions are fully methodological: adding the support of multilingualism to LLM during training and SFT phases. It is unavoidable that PolyLM might exhibit several common deficiencies of language models, e.g. hallucination and toxicity. PolyLM should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application.
BibTeX:
| [
"# Model Card for PolyLM-Multialpaca\n\nThis model is finetuned on polyLM-13b using the following datasets:",
"# Demo\nOpen",
"# Bias, Risks, and Limitations\n\nThe information below in this section are copied from the model's official model card:\n\n> Our contributions are fully methodological: adding the support of multilingualism to LLM during training and SFT phases. It is unavoidable that PolyLM might exhibit several common deficiencies of language models, e.g. hallucination and toxicity. PolyLM should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application.\n\nBibTeX:"
] | [
"TAGS\n#gguf #arxiv-2307.06018 #license-apache-2.0 #region-us \n",
"# Model Card for PolyLM-Multialpaca\n\nThis model is finetuned on polyLM-13b using the following datasets:",
"# Demo\nOpen",
"# Bias, Risks, and Limitations\n\nThe information below in this section are copied from the model's official model card:\n\n> Our contributions are fully methodological: adding the support of multilingualism to LLM during training and SFT phases. It is unavoidable that PolyLM might exhibit several common deficiencies of language models, e.g. hallucination and toxicity. PolyLM should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application.\n\nBibTeX:"
] | [
25,
29,
3,
122
] | [
"passage: TAGS\n#gguf #arxiv-2307.06018 #license-apache-2.0 #region-us \n# Model Card for PolyLM-Multialpaca\n\nThis model is finetuned on polyLM-13b using the following datasets:# Demo\nOpen# Bias, Risks, and Limitations\n\nThe information below in this section are copied from the model's official model card:\n\n> Our contributions are fully methodological: adding the support of multilingualism to LLM during training and SFT phases. It is unavoidable that PolyLM might exhibit several common deficiencies of language models, e.g. hallucination and toxicity. PolyLM should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application.\n\nBibTeX:"
] | [
-0.07694800198078156,
-0.003832208923995495,
-0.0032171246130019426,
0.08892849087715149,
0.09044913947582245,
0.02589600160717964,
0.18602341413497925,
-0.004531085956841707,
0.07030679285526276,
-0.019696708768606186,
0.15103453397750854,
0.020258476957678795,
0.043562617152929306,
0.09524308145046234,
0.01582084782421589,
-0.11507944017648697,
0.05005201697349548,
-0.052176568657159805,
0.012144417501986027,
0.08980162441730499,
0.06570253521203995,
-0.018752053380012512,
0.10131511837244034,
0.07999244332313538,
-0.12267062813043594,
-0.06206631660461426,
0.05898025259375572,
-0.005066779907792807,
0.06277142465114594,
0.15878790616989136,
0.008600753732025623,
0.026749469339847565,
-0.0034254705533385277,
-0.19390162825584412,
0.055648382753133774,
0.015892144292593002,
-0.021240461617708206,
0.04429454356431961,
0.1291772425174713,
0.015995807945728302,
0.08691176772117615,
0.06365051865577698,
-0.0063245659694075584,
0.017764870077371597,
-0.06842371076345444,
-0.10696760565042496,
-0.05762137845158577,
0.02865944802761078,
-0.02401825040578842,
-0.015763085335493088,
0.0115593820810318,
0.21071234345436096,
-0.04131484404206276,
0.07517969608306885,
0.1968301236629486,
-0.10830460488796234,
-0.03979380056262016,
0.19104230403900146,
0.012347005307674408,
0.0050826119258999825,
0.03246651962399483,
0.07963012903928757,
0.10748158395290375,
0.05019259452819824,
-0.06129361689090729,
-0.08392156660556793,
0.10019765049219131,
-0.03005678579211235,
-0.11097406595945358,
-0.0017286428483203053,
0.2261086404323578,
-0.008943716995418072,
-0.11488354951143265,
-0.04150596633553505,
-0.0069146528840065,
0.13851334154605865,
0.02983885630965233,
0.010637007653713226,
-0.021943241357803345,
0.0038618966937065125,
0.1414026916027069,
-0.050692640244960785,
-0.01388514507561922,
-0.0852537676692009,
-0.03241182491183281,
0.0918867215514183,
0.04362501576542854,
0.1072068139910698,
-0.01081516221165657,
0.042481038719415665,
-0.15592554211616516,
0.032105952501297,
-0.05459418520331383,
-0.17261767387390137,
0.026532676070928574,
0.02263651229441166,
-0.0006686863489449024,
0.06094110384583473,
0.11103308945894241,
0.011286973021924496,
0.024363987147808075,
0.06852773576974869,
-0.008620686829090118,
0.05564076825976372,
0.11172423511743546,
-0.09419254958629608,
-0.06949120759963989,
0.1165042594075203,
0.05829048156738281,
0.03141498565673828,
0.08675312250852585,
-0.06686113029718399,
-0.07664897292852402,
0.015816403552889824,
-0.04053659737110138,
0.12663088738918304,
-0.03533504158258438,
0.018443556502461433,
-0.06923719495534897,
-0.07633106410503387,
0.07392867654561996,
-0.029881982132792473,
-0.03926095366477966,
-0.03524526581168175,
-0.009471233002841473,
0.05583886057138443,
0.10729243606328964,
-0.07626442611217499,
-0.013946878723800182,
0.03358697518706322,
-0.07592379301786423,
0.07103990018367767,
-0.07475942373275757,
-0.053284648805856705,
-0.00030540963052771986,
-0.014104314148426056,
0.036847978830337524,
-0.09588226675987244,
-0.13205581903457642,
-0.021075395867228508,
0.04985851049423218,
-0.048299722373485565,
0.04925168678164482,
0.07811319082975388,
0.056225214153528214,
-0.10134409368038177,
-0.06336165964603424,
0.10880370438098907,
-0.09316936135292053,
0.04249073937535286,
-0.05640912801027298,
0.061732277274131775,
-0.1756436675786972,
0.016566701233386993,
-0.04863180220127106,
0.11224249750375748,
0.03606736287474632,
-0.014340603724122047,
-0.07476132363080978,
-0.005528680514544249,
-0.023272152990102768,
-0.11288566887378693,
0.04544326290488243,
0.05237707495689392,
-0.007100661750882864,
0.061749838292598724,
-0.13951022922992706,
0.001695266691967845,
0.18862318992614746,
-0.19207997620105743,
-0.10213334113359451,
0.09157522767782211,
-0.0688365176320076,
-0.007771856617182493,
0.03289870172739029,
0.03370046988129616,
0.05754440650343895,
-0.13640309870243073,
0.07259548455476761,
0.02276705391705036,
-0.05666718631982803,
0.014233374036848545,
0.0498448982834816,
0.042035192251205444,
-0.3154531717300415,
0.04153056815266609,
0.0004397814627736807,
0.050060153007507324,
-0.028852319344878197,
-0.039389293640851974,
-0.04022494703531265,
-0.06665214896202087,
0.029669376090168953,
0.0020589001942425966,
-0.05768008157610893,
-0.03337932750582695,
0.032951075583696365,
0.06544969230890274,
0.1174841895699501,
0.011881022714078426,
-0.03223568573594093,
-0.08193805068731308,
0.12764872610569,
-0.05238654837012291,
0.0020597216207534075,
-0.08978558331727982,
-0.04417646303772926,
-0.07140643894672394,
0.010491990484297276,
0.05651672184467316,
0.18190310895442963,
-0.008085019886493683,
0.1365656703710556,
-0.02889499068260193,
0.1313403993844986,
0.12336811423301697,
-0.003364787669852376,
-0.0029166063759475946,
-0.17443369328975677,
0.0547269769012928,
-0.03731725364923477,
0.11155648529529572,
-0.11987268924713135,
-0.016902176663279533,
-0.024404695257544518,
-0.05722062662243843,
-0.04246557876467705,
0.01264810562133789,
0.11861123889684677,
0.027014032006263733,
-0.0047672344371676445,
-0.04212239384651184,
0.08384736627340317,
-0.023243317380547523,
-0.01688419096171856,
0.04568878933787346,
-0.14998216927051544,
0.2296370267868042,
0.13864931464195251,
0.059015367180109024,
-0.04932574927806854,
-0.09821008890867233,
-0.010074890218675137,
0.031413961201906204,
-0.10649333894252777,
0.13613012433052063,
0.022854695096611977,
-0.02206721343100071,
0.08766794204711914,
-0.11886294186115265,
0.045312125235795975,
0.06515113264322281,
-0.16769129037857056,
-0.032112907618284225,
0.08093108236789703,
0.09715206176042557,
-0.12483306974172592,
0.0764046311378479,
-0.001975195249542594,
-0.048192836344242096,
0.14051494002342224,
0.07421095669269562,
0.03934433311223984,
-0.07237982749938965,
0.05937943235039711,
-0.020107094198465347,
0.23590245842933655,
-0.07219230383634567,
0.05399234592914581,
0.0658126249909401,
0.018018430098891258,
0.05957019329071045,
-0.15055759251117706,
-0.08229091763496399,
-0.04341927543282509,
-0.031060896813869476,
-0.16397863626480103,
0.03628945350646973,
-0.015453221276402473,
0.08706676214933395,
-0.05955834686756134,
-0.0068066855892539024,
0.06988118588924408,
0.002452797954902053,
-0.11955127120018005,
0.13241413235664368,
-0.13524387776851654,
-0.22637055814266205,
-0.12424709647893906,
-0.14316068589687347,
-0.16458700597286224,
0.017178192734718323,
0.018040381371974945,
-0.06593352556228638,
-0.0667400136590004,
-0.08010830730199814,
-0.028963550925254822,
-0.0940340980887413,
-0.03947169706225395,
-0.016934378072619438,
-0.0002702399797271937,
0.0708577111363411,
-0.06521103531122208,
0.005096806678920984,
-0.046287573873996735,
0.06651880592107773,
0.09926014393568039,
-0.08382144570350647,
0.043104927986860275,
0.16239884495735168,
0.009009637869894505,
0.014544835314154625,
0.002994925482198596,
0.23522348701953888,
0.011779049411416054,
-0.12249913066625595,
0.25456541776657104,
-0.06749080121517181,
0.030768046155571938,
0.040056996047496796,
0.07324011623859406,
-0.11513053625822067,
-0.005369724240154028,
0.01637379638850689,
-0.1690138429403305,
-0.19349227845668793,
-0.07475081831216812,
-0.007339324336498976,
0.00671524228528142,
-0.03759780526161194,
0.06162228435277939,
0.033764466643333435,
0.15501214563846588,
0.0005440223612822592,
0.0009377817041240633,
-0.035837966948747635,
0.023428859189152718,
0.11565066874027252,
-0.08134925365447998,
0.06900850683450699,
-0.07243205606937408,
0.004393525887280703,
0.156962051987648,
0.07119038701057434,
0.329332172870636,
0.022681133821606636,
-0.17607702314853668,
0.22090983390808105,
0.03306392580270767,
0.06637033075094223,
0.02203509397804737,
-0.05494624748826027,
-0.0171196348965168,
-0.07410869002342224,
-0.05677071586251259,
-0.10194867104291916,
0.05503814294934273,
-0.041549790650606155,
-0.009177686646580696,
-0.010093833319842815,
-0.011540220119059086,
0.02712361328303814,
-0.14940525591373444,
0.016683200374245644,
-0.14172856509685516,
0.0012491989182308316,
0.05496986210346222,
0.06882993131875992,
-0.11099718511104584,
0.05425036698579788,
0.08539269119501114,
-0.1013217493891716,
0.1562705785036087,
-0.048224981874227524,
0.06516426801681519,
-0.03191649168729782,
0.013665376231074333,
-0.1368866264820099,
0.08710886538028717,
-0.08198083937168121,
0.14966441690921783,
-0.34629103541374207,
0.17509053647518158,
0.016108987852931023,
-0.0648648589849472,
-0.118135005235672,
-0.017872612923383713,
0.09130501747131348,
0.19485437870025635,
0.13096259534358978,
0.06944487243890762,
-0.1526230126619339,
-0.20511095225811005,
-0.030163483694195747,
0.05201154574751854,
-0.0082355085760355,
0.22761215269565582,
-0.04340561851859093,
0.03356752544641495,
0.04979323968291283,
-0.031081920489668846,
-0.10168979316949844,
-0.1287735551595688,
-0.09490146487951279,
0.0861990749835968,
0.002885835012421012,
-0.17018286883831024,
-0.07362213730812073,
-0.05840589106082916,
-0.03338106721639633,
0.1291760802268982,
0.002934960648417473,
-0.11434105038642883,
-0.07231183350086212,
-0.042657241225242615,
0.03621693328022957,
0.0027412641793489456,
-0.003192390315234661,
-0.09245608001947403,
-0.008276990614831448,
-0.04749966040253639,
-0.09338672459125519,
0.05786750838160515,
-0.1455121636390686,
-0.08584006130695343,
-0.06835529208183289,
0.11968301981687546,
0.05790386348962784,
0.022741910070180893,
0.028080124408006668,
-0.07641354948282242,
-0.06068619340658188,
-0.2056291252374649,
0.015184956602752209,
0.09398236125707626,
0.007703106850385666,
0.07751878350973129,
0.08263856172561646,
0.01833842694759369,
-0.011872592382133007,
-0.09450102597475052,
0.023600338026881218,
0.2144935131072998,
-0.033418238162994385,
0.1514320820569992,
0.07256197929382324,
-0.0919070616364479,
-0.16133546829223633,
-0.0058693611063063145,
-0.03717556968331337,
-0.011039086617529392,
0.03803236782550812,
-0.12171925604343414,
0.0006854871171526611,
0.09518525749444962,
-0.016950054094195366,
0.1860760897397995,
-0.21523796021938324,
-0.08216924965381622,
0.18402840197086334,
0.1497962772846222,
0.43272271752357483,
-0.1094064936041832,
-0.05557661131024361,
-0.1125682145357132,
-0.03430655598640442,
0.1718471199274063,
-0.07245397567749023,
0.08355364203453064,
-0.03512527421116829,
0.042540293186903,
0.029350392520427704,
-0.015387076884508133,
0.2959645092487335,
-0.042531728744506836,
0.1354050636291504,
-0.08489198982715607,
0.010190414264798164,
-0.09291708469390869,
0.000767428136896342,
0.09427357465028763,
-0.04165279492735863,
0.016439266502857208,
0.022464340552687645,
-0.10069359093904495,
-0.03955268859863281,
0.047286879271268845,
-0.003386890282854438,
-0.08396410197019577,
-0.13438132405281067,
0.0649205893278122,
-0.05241767317056656,
-0.003219956997781992,
-0.15032759308815002,
-0.09052886068820953,
-0.02928691916167736,
0.1431577354669571,
0.12777645885944366,
-0.1760849505662918,
-0.07280638068914413,
0.02829105779528618,
-0.042088013142347336,
0.07331902533769608,
-0.11767878383398056,
-0.006036446895450354,
0.03545019403100014,
0.004468314815312624,
0.09931831061840057,
0.00880266260355711,
-0.12068426609039307,
0.04323720932006836,
0.10144593566656113,
-0.11571332812309265,
-0.11188877373933792,
-0.019977401942014694,
0.11855223029851913,
0.05666905269026756,
-0.013378257863223553,
0.08898837864398956,
-0.020071694627404213,
0.044585421681404114,
-0.018493253737688065,
0.049747440963983536,
-0.01957547664642334,
0.1197386234998703,
0.08829981833696365,
0.0360809862613678,
-0.10548536479473114,
-0.001935972017236054,
-0.005987323820590973,
0.12960337102413177,
0.010945268906652927,
-0.1620723158121109,
-0.07526306807994843,
-0.04059726744890213,
-0.15344315767288208,
0.03299643471837044,
-0.0025432980619370937,
-0.21985432505607605,
-0.038807764649391174,
-0.16262474656105042,
-0.06419266015291214,
0.05775560438632965,
0.07826223224401474,
0.03654199466109276,
-0.014171168208122253,
-0.08958276361227036,
-0.038677193224430084,
0.010550851002335548,
-0.012124762870371342,
0.022834770381450653,
-0.12235401570796967,
0.07624778151512146,
0.0970921441912651,
0.14254550635814667,
-0.08636511117219925,
-0.03026507794857025,
-0.12249494343996048,
0.0382537916302681,
-0.22844509780406952,
0.057666923850774765,
-0.1420580893754959,
0.05592074617743492,
0.028506474569439888,
0.0019516393076628447,
-0.02466536872088909,
0.052693676203489304,
-0.0720212534070015,
0.09289777278900146,
0.04397145286202431,
0.03397679701447487,
-0.06952307373285294,
-0.05257824808359146,
0.06306612491607666,
0.019796079024672508,
0.04109525680541992,
-0.01057344675064087,
-0.07838217169046402,
0.04794350638985634,
-0.1984194964170456,
0.032647475600242615,
0.09178347140550613,
0.056431811302900314,
-0.05588599294424057,
-0.22909443080425262,
0.045055750757455826,
0.0537266805768013,
-0.0306383166462183,
0.09188595414161682,
0.060936275869607925,
-0.07938896119594574,
-0.05414266511797905,
0.01928057335317135,
-0.044718023389577866,
0.05464351549744606,
0.0027072164230048656,
0.04310450330376625,
0.05847802013158798,
0.011829221621155739,
-0.020077962428331375,
0.013510327786207199,
-0.05579977482557297,
-0.023819053545594215,
-0.04447745159268379,
-0.061329569667577744,
0.027554618194699287,
-0.01649453490972519,
-0.025460388511419296,
0.057210177183151245,
0.430499792098999,
0.04881652444601059,
-0.05113552138209343,
-0.0462532564997673,
0.1272704154253006,
0.05212549865245819,
-0.0559835284948349,
0.04654036834836006,
-0.07490743696689606,
0.0823952853679657,
0.0015462142182514071,
0.06681032478809357,
0.05924622341990471,
-0.03580663353204727,
0.1680123656988144,
-0.03537093102931976,
-0.03634634613990784,
-0.005322868470102549,
0.01919514313340187,
-0.1156001091003418,
0.04933547601103783,
0.020586229860782623,
0.004700740799307823,
0.042637359350919724,
-0.0490388385951519,
-0.04517650976777077,
0.05533965304493904,
-0.04250435531139374,
0.0298274178057909,
0.013002228923141956,
-0.08878879994153976,
-0.1862887591123581,
-0.029183408245444298,
-0.021007617935538292,
-0.09965486079454422,
-0.011382819153368473,
-0.07818280160427094,
-0.0442439541220665,
0.2742438316345215,
0.023486021906137466,
-0.033523399382829666,
0.05316277965903282,
-0.15386976301670074,
-0.02642570249736309,
-0.08765165507793427,
-0.011895166710019112,
-0.02967662550508976,
-0.07484336197376251,
0.0048394170589745045,
0.057173844426870346,
-0.07673332840204239,
0.017868515104055405,
0.029162688180804253,
0.11974970251321793,
0.013576487079262733,
-0.0506722666323185,
0.010059686377644539,
-0.10777001082897186,
0.03726273775100708,
0.10914430767297745,
0.23014798760414124,
0.01858595572412014,
-0.03553197160363197,
-0.019684329628944397,
0.03413347154855728,
0.04198906198143959,
-0.03521297127008438,
-0.08883284777402878,
0.2615058124065399,
-0.15574851632118225,
-0.022345595061779022,
0.02950499951839447,
0.03203510865569115,
0.07929501682519913,
0.18731366097927094,
0.17291618883609772,
-0.1449279636144638,
-0.012744097039103508,
-0.07206686586141586,
0.015251676551997662,
0.0005691361147910357,
0.15557314455509186,
-0.036180444061756134,
0.14672277867794037,
-0.031093284487724304,
-0.03659401461482048,
-0.02218824438750744,
0.06448403000831604,
-0.04520600289106369,
0.009957813657820225,
0.08851027488708496,
-0.004721569363027811,
-0.056543298065662384,
0.06301639974117279,
-0.11482478678226471,
0.041354451328516006,
0.050549138337373734,
0.00034076947486028075,
-0.12834547460079193,
0.021486403420567513,
-0.08235359936952591,
-0.017391206696629524,
0.00598165113478899,
-0.13922904431819916,
0.03018365614116192,
0.1471380889415741,
0.014660179615020752,
-0.24082885682582855,
-0.10390552878379822,
0.12632381916046143,
0.22666330635547638,
0.1819060891866684,
0.04692931845784187,
0.20246319472789764,
0.03310727700591087,
-0.05379016697406769,
-0.050360050052404404,
0.1037902757525444,
-0.03706450015306473,
-0.019182298332452774,
-0.0136194359511137,
-0.06494098156690598,
0.025361936539411545,
-0.005433039739727974,
0.035139624029397964,
0.11616180837154388,
-0.02775949239730835,
0.024704743176698685,
-0.026210008189082146,
-0.046521980315446854,
0.09518205374479294,
-0.14782807230949402,
0.0746028795838356,
0.08145539462566376,
-0.04588286951184273,
-0.06245007365942001,
-0.04389912262558937,
0.12073260545730591,
-0.06342839449644089,
-0.06085095554590225,
0.02750742807984352,
-0.05704382434487343,
0.05578119680285454,
-0.06151438131928444,
-0.028625644743442535,
-0.28839099407196045,
-0.025227140635252,
-0.07605021446943283,
-0.012019415386021137,
0.003512469120323658,
0.0404927097260952,
-0.04602990299463272,
0.05127442628145218,
-0.04233638942241669,
-0.022357122972607613,
-0.03485402837395668,
-0.03956332430243492,
0.02134948968887329,
-0.06361009925603867
] |
null | null | transformers | # miqu-1-120b

* EXL2: [2.4bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2) | [2.65bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2) | [3.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2) | 4.0bpw | [5.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2)
* GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/) | [IQ3_XXS](https://huggingface.co/wolfram/miqu-1-120b-GGUF)
* HF FP16: [wolfram/miqu-1-120b](https://huggingface.co/wolfram/miqu-1-120b)
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
Inspired by [Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2), [MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [goliath-120b](https://huggingface.co/alpindale/goliath-120b).
Thanks for the support, [CopilotKit](https://github.com/CopilotKit/CopilotKit) - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, [Lone Striker](https://huggingface.co/LoneStriker)!
## Prompt template: Mistral
```
<s>[INST] {prompt} [/INST]
```
See also: [🐺🐦⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates : LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/18ljvxb/llm_prompt_format_comparisontest_mixtral_8x7b/)
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
dtype: float16
merge_method: passthrough
slices:
- sources:
- layer_range: [0, 20]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [10, 30]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [20, 40]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [30, 50]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [40, 60]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [50, 70]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [60, 80]
model: 152334H/miqu-1-70b-sf
```
## Credits & Special Thanks
* original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
* leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
* f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
* mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
* mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
### Support
* [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| {"language": ["en", "de", "fr", "es", "it"], "library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["152334H/miqu-1-70b-sf"]} | text-generation | LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"en",
"de",
"fr",
"es",
"it",
"base_model:152334H/miqu-1-70b-sf",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:46:49+00:00 | [] | [
"en",
"de",
"fr",
"es",
"it"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # miqu-1-120b
!image/jpeg
* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw
* GGUF: Q2_K-Q5_K_M | IQ3_XXS
* HF FP16: wolfram/miqu-1-120b
This is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.
Inspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.
Thanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, Lone Striker!
## Prompt template: Mistral
See also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* 152334H/miqu-1-70b-sf
### Configuration
The following YAML configuration was used to produce this model:
## Credits & Special Thanks
* original (unreleased) model: mistralai (Mistral AI_)
* leaked model: miqudev/miqu-1-70b
* f16 model: 152334H/miqu-1-70b-sf
* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.
* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2
### Support
* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| [
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
85,
206,
44,
31,
4,
17,
28,
17,
107,
69,
43
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140## Merge Details### Merge Method\n\nThis model was merged using the passthrough merge method.### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.08452296257019043,
0.07996358722448349,
-0.006216804962605238,
0.07489742338657379,
0.062447689473629,
0.048032015562057495,
0.10850238054990768,
0.11730548739433289,
0.04997159540653229,
0.1304655373096466,
0.019059622660279274,
0.0882313996553421,
0.10314993560314178,
0.1546645164489746,
-0.0008784975507296622,
-0.13700617849826813,
0.03406155854463577,
-0.032812073826789856,
-0.05265757441520691,
0.07489487528800964,
0.07002948969602585,
-0.049538981169462204,
0.07827795296907425,
0.01992727816104889,
-0.0794241651892662,
-0.02351442351937294,
-0.03350690379738808,
-0.0034194772597402334,
0.06382977962493896,
0.09272785484790802,
0.010276054963469505,
0.01770416460931301,
0.010751184076070786,
-0.1683615893125534,
0.032938648015260696,
0.03592948243021965,
-0.017932593822479248,
0.07718798518180847,
0.10788709670305252,
-0.018908003345131874,
0.021088771522045135,
-0.10027661174535751,
-0.02568136155605316,
0.08249933272600174,
-0.10719147324562073,
-0.1474289745092392,
-0.18762712180614471,
0.15491966903209686,
0.042507048696279526,
0.03724713996052742,
-0.011902958154678345,
0.1333865076303482,
0.050249699503183365,
0.06016452983021736,
0.24593202769756317,
-0.22131793200969696,
-0.061919860541820526,
0.028957664966583252,
0.05541115999221802,
-0.04985284432768822,
-0.021356018260121346,
0.01992829702794552,
0.04054830223321915,
0.02191905304789543,
-0.020568734034895897,
-0.04234932363033295,
0.13632027804851532,
-0.05843184515833855,
-0.13171182572841644,
-0.02820662222802639,
0.09477535635232925,
0.04355064406991005,
-0.060202740132808685,
-0.1276080310344696,
-0.0682656466960907,
-0.03270931541919708,
-0.0022935839369893074,
-0.04131745919585228,
-0.002290143631398678,
-0.011816466227173805,
0.08523242175579071,
-0.08333887904882431,
-0.05337991565465927,
-0.020235886797308922,
-0.06682160496711731,
0.16135647892951965,
0.018354052677750587,
0.013679140247404575,
-0.013255320489406586,
0.07744071632623672,
-0.1265559196472168,
-0.1336318850517273,
-0.05428304150700569,
-0.03462158143520355,
-0.05692019686102867,
-0.05218517407774925,
0.01661340333521366,
-0.07750479131937027,
0.08107020705938339,
0.175857275724411,
-0.11741548776626587,
0.06220667064189911,
-0.009697522968053818,
0.030903007835149765,
0.024629943072795868,
0.08424429595470428,
-0.08733733743429184,
-0.08122188597917557,
0.04335138201713562,
0.05078619346022606,
0.05994807183742523,
0.0019238290842622519,
0.00010135882621398196,
-0.02575274370610714,
0.0019080113852396607,
0.07156599313020706,
0.06952427327632904,
0.029285242781043053,
-0.07679720222949982,
-0.05144278332591057,
0.17693321406841278,
-0.10877041518688202,
0.05875063315033913,
0.038095518946647644,
-0.010931058786809444,
-0.015534471720457077,
0.026578398421406746,
-0.028613878414034843,
-0.05422482639551163,
0.014574683271348476,
-0.052837468683719635,
0.006087095942348242,
-0.04121243208646774,
-0.04512706398963928,
0.052960656583309174,
0.044456131756305695,
-0.05848181992769241,
-0.12249575555324554,
-0.08548451960086823,
-0.056048691272735596,
0.033982738852500916,
-0.07165446132421494,
0.009947382844984531,
0.03666377440094948,
-0.08962082862854004,
0.029198521748185158,
0.02603435143828392,
0.022042686119675636,
-0.022196292877197266,
0.013299040496349335,
0.01655365526676178,
0.02530047670006752,
-0.031476203352212906,
-0.009267413057386875,
-0.05801313742995262,
0.08978775143623352,
-0.19187140464782715,
0.07319818437099457,
-0.06753759831190109,
-0.0458916574716568,
-0.1367833912372589,
-0.022929007187485695,
0.0001557070208946243,
-0.021685365587472916,
0.08853033930063248,
0.14028236269950867,
-0.15254990756511688,
-0.01800709031522274,
0.133183553814888,
-0.09939759224653244,
-0.08947071433067322,
0.09990517795085907,
0.010765177197754383,
-0.005259351339191198,
0.010271748527884483,
0.08585681021213531,
0.12536770105361938,
-0.056073881685733795,
-0.0745566338300705,
-0.10543492436408997,
-0.02353603020310402,
0.10377749055624008,
0.036487940698862076,
-0.06648851186037064,
0.006608173716813326,
0.0013413212727755308,
-0.03926616534590721,
-0.02470051497220993,
-0.01713505946099758,
-0.04516416788101196,
-0.051653604954481125,
-0.03216031193733215,
0.019755378365516663,
-0.0311356820166111,
-0.04306983947753906,
-0.06874691694974899,
-0.0662546455860138,
-0.03937284275889397,
0.10621660202741623,
0.0002958594122901559,
0.013465981930494308,
-0.06993132084608078,
0.12132570147514343,
-0.043282829225063324,
0.019509179517626762,
-0.13229045271873474,
-0.062326252460479736,
0.05427157133817673,
-0.17642687261104584,
0.057915039360523224,
-0.07802051305770874,
0.07007356733083725,
0.08792263269424438,
-0.03142979368567467,
-0.04443249851465225,
0.014312747865915298,
-0.011503690853714943,
-0.05503445118665695,
-0.13744132220745087,
-0.05339666083455086,
-0.03718583658337593,
0.1360798180103302,
-0.07717297226190567,
0.03823299705982208,
0.04849797114729881,
0.17298123240470886,
0.013948954641819,
-0.06230805069208145,
0.059276606887578964,
0.013416923582553864,
-0.01901276968419552,
-0.04254152253270149,
0.022224929183721542,
-0.015195275656878948,
-0.06335578858852386,
0.06797880679368973,
-0.16682805120944977,
-0.09226492792367935,
0.08205856382846832,
0.11543388664722443,
-0.10825490206480026,
-0.0011853284668177366,
-0.0016989412251859903,
-0.03144387900829315,
-0.02349192649126053,
-0.07283974438905716,
0.139337420463562,
0.03313908725976944,
0.05388158559799194,
-0.05538925528526306,
-0.07884406298398972,
0.012225146405398846,
-0.024691954255104065,
-0.07743193954229355,
0.12690365314483643,
0.02291349321603775,
-0.19036413729190826,
0.1086549460887909,
0.10671419650316238,
0.024346476420760155,
0.12972186505794525,
0.010163333266973495,
-0.020192109048366547,
-0.11323220282793045,
0.07214321941137314,
0.06328365206718445,
0.08764441311359406,
0.009260048158466816,
0.06931837648153305,
0.03807743266224861,
0.00207220995798707,
0.013925828970968723,
-0.11039725691080093,
0.03326145187020302,
0.019972439855337143,
-0.035299137234687805,
0.07873526960611343,
0.04182936251163483,
0.027880189940333366,
0.08705969154834747,
0.017179058864712715,
0.07423700392246246,
-0.01186748594045639,
-0.03507009521126747,
-0.09978814423084259,
0.13680647313594818,
-0.10421305149793625,
-0.1235826313495636,
-0.15689323842525482,
0.023495908826589584,
-0.1045532375574112,
-0.031237754970788956,
0.04314960539340973,
-0.04964744672179222,
-0.04190392047166824,
-0.09863945841789246,
0.07575149089097977,
0.04664899781346321,
-0.031183557584881783,
0.029061678797006607,
-0.04268911853432655,
0.05956493318080902,
-0.1079893708229065,
-0.03650837019085884,
0.0311584435403347,
0.003986986353993416,
0.03826166316866875,
0.06534901261329651,
0.09445244818925858,
0.06816873699426651,
0.010970757342875004,
-0.007111722603440285,
0.0249604694545269,
0.2451404184103012,
-0.09102492034435272,
0.08363116532564163,
0.1731153130531311,
0.012367261573672295,
0.08963974565267563,
0.1418784111738205,
0.0519220232963562,
-0.06685937941074371,
-0.013835576362907887,
0.023638667538762093,
-0.024145755916833878,
-0.19078323245048523,
-0.061981018632650375,
-0.049455784261226654,
0.02179974503815174,
0.016594912856817245,
0.056652046740055084,
-0.03246590495109558,
0.04957956075668335,
-0.06632956862449646,
-0.023145724087953568,
0.04272812604904175,
0.07885605841875076,
0.04942423850297928,
0.05232582241296768,
0.032079048454761505,
-0.030630147084593773,
-0.01874661073088646,
0.10800144821405411,
0.002024435205385089,
0.0828050896525383,
0.005763496737927198,
0.15540826320648193,
0.042188387364149094,
0.05763785168528557,
-0.006694222800433636,
0.05031910166144371,
0.012587900273501873,
0.01727849431335926,
0.011400274932384491,
-0.0922514945268631,
0.017045721411705017,
0.037224024534225464,
0.031022431328892708,
0.08605832606554031,
-0.019477766007184982,
-0.042653512209653854,
0.06828916072845459,
0.20636118948459625,
0.054764457046985626,
-0.18104644119739532,
-0.052608687430620193,
0.05323893949389458,
-0.005172972101718187,
-0.0441189780831337,
-0.0574423223733902,
0.0601525716483593,
-0.1421785205602646,
0.10297753661870956,
-0.02837306633591652,
0.08344317972660065,
-0.0768112912774086,
-0.025202374905347824,
-0.002551008015871048,
0.14314106106758118,
-0.007685830816626549,
0.009870217181742191,
-0.11473074555397034,
0.09677925705909729,
0.047673650085926056,
0.050363969057798386,
-0.0467650331556797,
0.05014784261584282,
0.06751372665166855,
-0.016121957451105118,
0.1009698212146759,
0.015580941922962666,
-0.08597638458013535,
-0.13121242821216583,
-0.07546716183423996,
-0.02328302152454853,
0.09724711626768112,
-0.07956860214471817,
0.11797893047332764,
-0.006677214056253433,
-0.041914381086826324,
-0.051285892724990845,
0.08619783818721771,
-0.16726025938987732,
-0.10657192021608353,
0.08049735426902771,
0.035933516919612885,
-0.00293326354585588,
-0.07889305055141449,
-0.03560107201337814,
-0.06792481988668442,
0.20780132710933685,
-0.12493043392896652,
-0.06458897888660431,
-0.12241725623607635,
-0.02887725457549095,
0.14838698506355286,
-0.09519194811582565,
0.016095273196697235,
-0.0620892271399498,
0.09675031155347824,
-0.05750473216176033,
-0.09322291612625122,
0.03633593022823334,
-0.0772782415151596,
-0.22163733839988708,
-0.029905345290899277,
0.1838245987892151,
0.031322285532951355,
0.03489852324128151,
0.0206821970641613,
0.04595774784684181,
0.021382097154855728,
-0.10996323078870773,
0.04105592891573906,
0.12817256152629852,
0.020900849252939224,
0.09039270132780075,
-0.045222893357276917,
-0.14004118740558624,
-0.09013182669878006,
-0.05677556246519089,
0.0997968390583992,
0.29861730337142944,
-0.05223260074853897,
0.030688846483826637,
0.04147891700267792,
-0.07043230533599854,
-0.15785279870033264,
-0.038915667682886124,
0.04657349735498428,
0.01753225363790989,
0.01858132891356945,
-0.07110119611024857,
0.0366257019340992,
0.10373353213071823,
-0.01873767003417015,
0.15667074918746948,
-0.263369619846344,
-0.12704867124557495,
-0.006682985462248325,
0.06700930744409561,
-0.01957916095852852,
-0.1452975571155548,
-0.11674544960260391,
-0.03796976059675217,
-0.1995268017053604,
0.09981389343738556,
-0.02467156946659088,
0.0965239554643631,
-0.023898916319012642,
0.015778150409460068,
0.03433018922805786,
-0.059150923043489456,
0.19815395772457123,
-0.039659541100263596,
0.04237253591418266,
-0.08535932749509811,
0.011777906678617,
0.026407601311802864,
-0.05703800171613693,
0.0950261577963829,
-0.035693515092134476,
0.017066331580281258,
-0.07961391657590866,
0.0030619590543210506,
-0.08434901386499405,
0.06988205015659332,
-0.053291499614715576,
-0.002292127348482609,
-0.06203543022274971,
0.07322024554014206,
0.02764309197664261,
-0.02415998838841915,
0.024202430620789528,
-0.021172653883695602,
0.10087262094020844,
0.2350863218307495,
0.09775305539369583,
0.05048970505595207,
-0.06141071021556854,
-0.00026774543221108615,
-0.04018312692642212,
0.02579716220498085,
-0.035933900624513626,
0.04462524875998497,
0.08787034451961517,
0.02455063909292221,
0.10782863199710846,
0.01810472458600998,
-0.1601009964942932,
0.018065448850393295,
0.06646080315113068,
-0.13265873491764069,
-0.2303391844034195,
0.003492377931252122,
0.04154038056731224,
-0.05837193876504898,
0.01806606538593769,
0.1862204670906067,
0.0009367137681692839,
-0.04963487386703491,
0.022825637832283974,
0.05416732281446457,
-0.03536570817232132,
0.15667866170406342,
0.00837104581296444,
0.04538920149207115,
-0.061773963272571564,
0.07580430060625076,
0.04042122885584831,
0.0016077554319053888,
0.018714869394898415,
0.17067234218120575,
-0.05190492421388626,
-0.07857772707939148,
0.04102415591478348,
0.03600002080202103,
-0.021241383627057076,
-0.021828800439834595,
-0.027304736897349358,
-0.10624261945486069,
0.013062734156847,
0.12654060125350952,
0.02681851200759411,
-0.018991822376847267,
0.08323623985052109,
-0.004140008706599474,
-0.028763674199581146,
0.07244459539651871,
0.05491362512111664,
0.08497325330972672,
-0.059819940477609634,
0.06284529715776443,
-0.02697847969830036,
0.03915136307477951,
-0.0026654857210814953,
0.019256316125392914,
-0.12592482566833496,
-0.05442740023136139,
-0.08961838483810425,
0.005896841175854206,
-0.11454344540834427,
-0.029031215235590935,
-0.0207052081823349,
-0.00936585757881403,
-0.02115669846534729,
0.03387326002120972,
-0.051367443054914474,
-0.11562348157167435,
-0.07470326870679855,
0.08342844247817993,
-0.14418748021125793,
-0.01641484536230564,
0.06919168680906296,
-0.07516539096832275,
0.09250383824110031,
0.01708120107650757,
0.0034909669775515795,
-0.009268661960959435,
-0.11126343160867691,
-0.02617836929857731,
-0.013749503530561924,
0.020847713574767113,
0.02727784588932991,
-0.20117603242397308,
-0.002813884522765875,
-0.04270019009709358,
-0.03368119150400162,
0.01241485308855772,
0.15425018966197968,
-0.09736622869968414,
-0.040856342762708664,
-0.03006616234779358,
-0.044825147837400436,
-0.06249745562672615,
0.018799860030412674,
0.10122360289096832,
0.009257921017706394,
0.08748460561037064,
-0.05421772226691246,
0.07078760117292404,
-0.17954121530056,
-0.01305952575057745,
0.022268623113632202,
-0.06352261453866959,
0.042893052101135254,
0.006506424397230148,
0.07432542741298676,
-0.024414660409092903,
0.033989157527685165,
-0.08212705701589584,
-0.018906600773334503,
0.026504697278141975,
-0.028520310297608376,
-0.021283207461237907,
0.017214683815836906,
0.020455483347177505,
-0.022489050403237343,
-0.024124853312969208,
-0.0313778817653656,
0.020932979881763458,
-0.019497908651828766,
-0.05568772181868553,
0.1307651400566101,
0.10116901993751526,
0.042394958436489105,
0.06033344939351082,
0.0699099600315094,
-0.06935380399227142,
-0.019917812198400497,
-0.015662092715501785,
-0.0501786470413208,
0.05921950191259384,
-0.0570748895406723,
0.11598431318998337,
0.0971740260720253,
-0.16128812730312347,
0.08817395567893982,
-0.03033575415611267,
-0.025268414989113808,
-0.0654015764594078,
-0.1747601330280304,
-0.06647072732448578,
-0.021373050287365913,
-0.007110625971108675,
-0.07304833084344864,
0.04277768358588219,
0.03683268278837204,
0.0017412106972187757,
-0.0061371480114758015,
0.12685923278331757,
-0.04393889009952545,
-0.058910902589559555,
0.035390980541706085,
0.037289973348379135,
-0.015837162733078003,
0.04472239688038826,
-0.006923268549144268,
0.022451741620898247,
0.012978330254554749,
0.05661128833889961,
0.08223818242549896,
-0.00805328506976366,
0.02584826387465,
-0.028271563351154327,
-0.10848145186901093,
-0.0006025367183610797,
0.005287060514092445,
0.01833254098892212,
0.08337931334972382,
0.05019523203372955,
0.016249405220150948,
-0.015350534580647945,
0.0810377299785614,
-0.02552863396704197,
-0.014691580086946487,
-0.1009829118847847,
0.15841826796531677,
-0.0559045635163784,
0.002254345454275608,
-0.0015080315060913563,
-0.11343378573656082,
-0.011597293429076672,
0.12920117378234863,
0.16337698698043823,
-0.08000685274600983,
0.03376281261444092,
0.017094949260354042,
0.00689410325139761,
0.004397694021463394,
0.049244899302721024,
0.059573933482170105,
0.10981784015893936,
-0.046494387090206146,
0.12055498361587524,
-0.0517595037817955,
-0.032632842659950256,
-0.07947799563407898,
0.09397511184215546,
-0.047008223831653595,
0.014640634879469872,
-0.017156826332211494,
0.07035721093416214,
-0.032913386821746826,
-0.16172361373901367,
0.03033449873328209,
-0.12312323600053787,
-0.138545423746109,
-0.00011208858632016927,
-0.0077992831356823444,
0.001197756384499371,
0.07809285819530487,
0.009989638812839985,
0.010767842642962933,
0.13262617588043213,
0.004294542595744133,
-0.07419104129076004,
-0.056784018874168396,
0.0005215644487179816,
-0.09722831845283508,
0.18984216451644897,
0.024254996329545975,
0.04178498685359955,
0.13091936707496643,
-0.033293675631284714,
-0.15407414734363556,
0.05585271492600441,
0.0700296014547348,
-0.09110260009765625,
0.022162333130836487,
0.14707012474536896,
-0.005258307326585054,
0.030615337193012238,
0.07387614995241165,
-0.053392838686704636,
0.019069518893957138,
0.0722138062119484,
0.019377747550606728,
-0.10239636152982712,
0.08977871388196945,
-0.10703950375318527,
0.13002605736255646,
0.18838851153850555,
-0.03489691764116287,
0.012865101918578148,
-0.03789073973894119,
0.04828639328479767,
0.038723692297935486,
0.09065895527601242,
-0.034608691930770874,
-0.19317959249019623,
0.06994378566741943,
-0.023887865245342255,
0.08475272357463837,
-0.1766246110200882,
-0.1239243745803833,
-0.029406068846583366,
-0.021863967180252075,
-0.013726417906582355,
0.0984341949224472,
0.09909792244434357,
0.0073255919851362705,
-0.023081263527274132,
-0.1434156447649002,
-0.02458539977669716,
0.10146303474903107,
-0.08631070703268051,
-0.04105154797434807
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - AndyMacool/youzi_lora
<Gallery />
## Model description
These are AndyMacool/youzi_lora LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using [DreamBooth](https://dreambooth.github.io/).
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use YOU12ZI to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](AndyMacool/youzi_lora/tree/main) them in the Files & versions tab.
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "YOU12ZI"} | text-to-image | AndyMacool/youzi_lora | [
"diffusers",
"tensorboard",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-06T06:50:32+00:00 | [] | [] | TAGS
#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - AndyMacool/youzi_lora
<Gallery />
## Model description
These are AndyMacool/youzi_lora LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use YOU12ZI to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
| [
"# SDXL LoRA DreamBooth - AndyMacool/youzi_lora\n\n<Gallery />",
"## Model description\n\nThese are AndyMacool/youzi_lora LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use YOU12ZI to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
"TAGS\n#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - AndyMacool/youzi_lora\n\n<Gallery />",
"## Model description\n\nThese are AndyMacool/youzi_lora LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use YOU12ZI to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
86,
23,
88,
16,
28
] | [
"passage: TAGS\n#diffusers #tensorboard #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - AndyMacool/youzi_lora\n\n<Gallery />## Model description\n\nThese are AndyMacool/youzi_lora LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.## Trigger words\n\nYou should use YOU12ZI to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
-0.08291801810264587,
0.15645891427993774,
0.0002142619778169319,
0.04682808369398117,
0.1519443839788437,
0.03818978741765022,
0.1825568526983261,
0.08811298757791519,
0.09585180878639221,
0.08578862994909286,
0.04177426919341087,
0.04136626049876213,
0.05594373866915703,
0.20016299188137054,
-0.018728015944361687,
-0.196250319480896,
0.011166650801897049,
0.0047994754277169704,
-0.04334042966365814,
0.023908259347081184,
0.04029730334877968,
-0.0775936171412468,
0.08661036193370819,
-0.024593690410256386,
-0.06058221682906151,
0.04006580263376236,
0.002358958823606372,
-0.030370088294148445,
0.013182793743908405,
0.053841885179281235,
0.06819406896829605,
0.06361761689186096,
0.07752515375614166,
-0.15658703446388245,
0.03212793543934822,
0.08959480375051498,
-0.02726241759955883,
0.05588458478450775,
-0.031047601252794266,
-0.09326529502868652,
0.07989686727523804,
-0.07435263693332672,
0.05213429406285286,
0.05927977338433266,
-0.023947784677147865,
-0.16884757578372955,
-0.030538314953446388,
0.005942912306636572,
0.051778268069028854,
0.047116901725530624,
0.011892176233232021,
0.061416905373334885,
0.06936758756637573,
0.03108985535800457,
0.23415707051753998,
-0.17842409014701843,
-0.06055667996406555,
0.26565366983413696,
0.04236364737153053,
0.12266526371240616,
-0.04420644789934158,
0.07231996953487396,
0.11616215109825134,
-0.012549338862299919,
0.10458153486251831,
-0.04695644974708557,
0.04995119944214821,
-0.07753933221101761,
-0.09884923696517944,
0.03994296118617058,
0.20782333612442017,
0.003401033114641905,
-0.07837113738059998,
-0.1530473530292511,
-0.06051810458302498,
0.08550504595041275,
-0.03386903554201126,
0.006842415314167738,
0.038366686552762985,
0.00647007068619132,
0.0016030512051656842,
-0.13664716482162476,
-0.041124314069747925,
-0.09646935015916824,
0.04436063393950462,
0.21095509827136993,
0.003698995104059577,
0.03578706085681915,
0.036105282604694366,
0.10904684662818909,
-0.1363627165555954,
-0.14281731843948364,
-0.004416401498019695,
-0.05212712660431862,
-0.012914354912936687,
0.022724885493516922,
0.010637586936354637,
-0.1722293347120285,
0.05132026970386505,
0.046552594751119614,
0.0247208084911108,
0.022949015721678734,
-0.04289531707763672,
0.039645005017519,
-0.012060137465596199,
0.08613928407430649,
-0.04171598330140114,
-0.09076439589262009,
0.05027511715888977,
0.09132847189903259,
0.09632711857557297,
-0.042764659970998764,
-0.12281142920255661,
-0.030442655086517334,
-0.10830000787973404,
0.07186061888933182,
0.01332944631576538,
-0.0037119321059435606,
-0.0602240189909935,
-0.06468245387077332,
0.1589227020740509,
-0.05613762512803078,
0.004444777499884367,
-0.03364325314760208,
-0.0661863163113594,
0.12980040907859802,
0.14708201587200165,
-0.0009438474080525339,
0.002922473940998316,
0.05074777081608772,
-0.07948119193315506,
0.005047009792178869,
-0.08149395138025284,
-0.11601120233535767,
0.011538354679942131,
-0.16316132247447968,
0.010267355479300022,
-0.10516814887523651,
-0.2590811550617218,
-0.040943533182144165,
0.040867604315280914,
-0.01850644126534462,
0.02660430781543255,
-0.05674157664179802,
-0.05377810820937157,
-0.01922285184264183,
0.03200030326843262,
-0.007723697926849127,
-0.009138672612607479,
0.051577139645814896,
-0.03255704045295715,
0.09260699152946472,
-0.03408588841557503,
0.02251097932457924,
-0.0747196152806282,
0.05380907282233238,
-0.16790954768657684,
0.1002480760216713,
-0.1110324114561081,
0.04424605891108513,
-0.09337332844734192,
-0.037598129361867905,
-0.024849584326148033,
0.009110819548368454,
0.025730833411216736,
0.13442924618721008,
-0.22235332429409027,
-0.05115899071097374,
0.13495057821273804,
-0.19476349651813507,
-0.08630642294883728,
0.022605378180742264,
-0.01794084534049034,
0.09023867547512054,
0.08505919575691223,
0.09426392614841461,
0.1837448626756668,
-0.25195756554603577,
-0.058991990983486176,
0.010960075072944164,
0.0026976580265909433,
-0.07141125947237015,
0.05454786866903305,
0.015164945274591446,
-0.04826219752430916,
0.051429834216833115,
-0.1415473073720932,
0.07818803191184998,
-0.004362261388450861,
-0.02677639201283455,
-0.048719361424446106,
-0.13492529094219208,
0.05822361633181572,
0.03261920064687729,
0.0062533230520784855,
0.021914368495345116,
-0.051984574645757675,
0.09224605560302734,
0.13963177800178528,
-0.09816915541887283,
-0.01165023259818554,
-0.0043583097867667675,
0.10902915149927139,
-0.16926918923854828,
-0.012612876482307911,
-0.06938280165195465,
-0.132393479347229,
0.058638639748096466,
0.008599256165325642,
0.018362291157245636,
-0.007228241767734289,
0.08790252357721329,
0.09702826291322708,
-0.06737730652093887,
-0.05511685833334923,
0.023191077634692192,
-0.01591479405760765,
-0.00741626275703311,
-0.10504470765590668,
-0.044067706912755966,
-0.08921607583761215,
0.08236090838909149,
-0.21020658314228058,
0.06957609951496124,
-0.013342389836907387,
0.10284847021102905,
0.10180837661027908,
-0.05984368547797203,
0.07991132885217667,
-0.020649755373597145,
-0.021666863933205605,
-0.10655619949102402,
-0.013383799232542515,
0.011143198236823082,
-0.14834840595722198,
0.08329103887081146,
-0.16442684829235077,
0.0499102883040905,
0.12445426732301712,
0.1716000884771347,
-0.014418507926166058,
-0.10116536915302277,
0.0038874177262187004,
-0.008199628442525864,
-0.09718693047761917,
-0.023611770942807198,
0.02361985668540001,
0.008437114767730236,
0.11100298166275024,
-0.07474876195192337,
0.04947388172149658,
0.03341294825077057,
-0.0062114764004945755,
-0.05592501536011696,
0.04792260378599167,
0.029960982501506805,
0.07235225290060043,
0.06289990246295929,
0.058198947459459305,
-0.0797748863697052,
0.19396886229515076,
0.0067456611432135105,
-0.06114239990711212,
-0.017391448840498924,
0.0029499358497560024,
0.05459267646074295,
0.10638336837291718,
0.08095892518758774,
0.055236175656318665,
0.016205791383981705,
-0.032957687973976135,
0.0027972888201475143,
-0.09432060271501541,
-0.03272569924592972,
0.02380499430000782,
-0.09367690235376358,
0.11786724627017975,
0.06366201490163803,
-0.07440252602100372,
0.03319453075528145,
-0.06832382082939148,
0.0747629925608635,
0.008190182037651539,
-0.022002525627613068,
-0.07952453941106796,
0.09681405127048492,
-0.061067551374435425,
-0.07965883612632751,
-0.15178510546684265,
0.059445131570100784,
-0.05966271460056305,
0.020719170570373535,
0.017002234235405922,
-0.04145744442939758,
-0.06779974699020386,
-0.10157206654548645,
0.04863913357257843,
0.03142755106091499,
-0.000442637741798535,
0.004217099864035845,
0.00364378048107028,
0.027420643717050552,
-0.11217048019170761,
0.008675169199705124,
-0.02599203959107399,
-0.04890231043100357,
-0.03146632760763168,
-0.019076796248555183,
0.1031981036067009,
0.10989821702241898,
0.014874408952891827,
0.028097033500671387,
-0.004028711933642626,
0.12829908728599548,
-0.04653934761881828,
0.11509864032268524,
0.26002341508865356,
0.10478558391332626,
0.06789790838956833,
0.08294045180082321,
0.02875586599111557,
-0.04096557945013046,
0.052824415266513824,
0.01809593290090561,
-0.14570724964141846,
-0.14522504806518555,
-0.12089400738477707,
-0.02874108962714672,
-0.015291307121515274,
0.09917635470628738,
0.0478619746863842,
0.09886237978935242,
0.16460424661636353,
-0.02510203793644905,
0.03571635112166405,
0.043120335787534714,
0.07887200266122818,
-0.05134505778551102,
-0.0014043851988390088,
0.041712746024131775,
-0.08770973235368729,
0.010625992901623249,
0.10273557901382446,
0.01966964453458786,
0.15707173943519592,
-0.10899801552295685,
0.04269206523895264,
0.004554584622383118,
0.025073185563087463,
0.06348072737455368,
0.08281911909580231,
-0.012160922400653362,
-0.024114754050970078,
-0.009325084276497364,
-0.12666325271129608,
0.03326920419931412,
0.12885193526744843,
-0.0001556959905428812,
0.07028041779994965,
-0.005359155125916004,
0.04312838613986969,
0.008190854452550411,
0.058954447507858276,
0.04197300598025322,
-0.28943175077438354,
-0.01638825424015522,
0.07617316395044327,
0.1139860674738884,
-0.029777424409985542,
0.007393205538392067,
0.18931183218955994,
-0.04941591992974281,
0.10335544496774673,
-0.05799616500735283,
0.06441488116979599,
-0.018552405759692192,
-0.07990612834692001,
0.0033025345765054226,
0.1951848268508911,
-0.03778162598609924,
0.016947804018855095,
-0.1571451723575592,
0.0029866553377360106,
0.028767697513103485,
0.1042831689119339,
-0.051838453859090805,
0.014392823912203312,
0.07969971001148224,
0.0347180999815464,
0.13861480355262756,
-0.01605582796037197,
0.039215803146362305,
-0.07013286650180817,
-0.13082465529441833,
-0.008198666386306286,
-0.008842702023684978,
-0.0925557017326355,
0.05103272944688797,
0.006914481054991484,
-0.015051872469484806,
0.010912735015153885,
-0.01798645220696926,
-0.1270606815814972,
-0.11101261526346207,
-0.03478770703077316,
0.15244249999523163,
0.005866712424904108,
-0.07339077442884445,
-0.08803065866231918,
-0.02868332900106907,
0.08002132922410965,
-0.03616243973374367,
-0.1392173320055008,
-0.1199447512626648,
-0.029949083924293518,
0.12721583247184753,
-0.05063223838806152,
0.016721252351999283,
-0.006052626296877861,
0.15301167964935303,
-0.12133818119764328,
-0.10986471176147461,
0.004268213175237179,
-0.07176350057125092,
-0.1358080357313156,
-0.03421488404273987,
0.09379816800355911,
0.01771928369998932,
0.008399679325520992,
-0.020292839035391808,
0.011889978311955929,
0.040761739015579224,
-0.11106526106595993,
0.03039722889661789,
0.22606821358203888,
-0.06688123196363449,
0.12286214530467987,
-0.01358508225530386,
-0.08840463310480118,
-0.005167635623365641,
0.0347982756793499,
0.00722303194925189,
0.22093366086483002,
-0.0892718955874443,
0.045094773173332214,
0.09422308951616287,
-0.06963907927274704,
-0.20498208701610565,
0.02807610109448433,
0.013100842945277691,
0.01801343448460102,
0.03265322744846344,
-0.16709847748279572,
0.13521036505699158,
0.04590187966823578,
-0.04524441063404083,
0.1959018111228943,
-0.3262695074081421,
-0.1258348971605301,
-0.007608386222273111,
0.15656106173992157,
0.14299853146076202,
-0.14615491032600403,
-0.06280476599931717,
-0.053021062165498734,
-0.02728632651269436,
0.13236266374588013,
-0.14503520727157593,
0.07327748835086823,
-0.006684525869786739,
-0.019561465829610825,
0.03879963979125023,
-0.06166968494653702,
0.12107128649950027,
-0.018845973536372185,
0.07637619227170944,
-0.029911741614341736,
0.011469793505966663,
0.10819923132658005,
-0.08625604212284088,
0.13686929643154144,
-0.15942876040935516,
0.05396832898259163,
-0.045216742902994156,
-0.05143129453063011,
-0.015064169652760029,
0.08169000595808029,
-0.01708425022661686,
-0.09623169898986816,
-0.04629966989159584,
0.04852529615163803,
0.052850931882858276,
0.009795550256967545,
-0.0031238803640007973,
-0.02468675933778286,
0.046681493520736694,
0.2092668116092682,
-0.0017277216538786888,
0.06660909205675125,
-0.05118144676089287,
0.0015701750526204705,
-0.044620219618082047,
0.10615228116512299,
-0.19146384298801422,
-0.0017022500978782773,
0.07844704389572144,
0.0596330463886261,
0.10410187393426895,
0.012656005099415779,
-0.10330428928136826,
0.06973662972450256,
0.04727704077959061,
-0.09624434262514114,
-0.13565930724143982,
-0.059600573033094406,
-0.05862155556678772,
-0.04450032487511635,
-0.0052266488783061504,
0.14389832317829132,
-0.09491445869207382,
0.03647098317742348,
-0.010973016731441021,
0.017794089391827583,
-0.026025056838989258,
0.09673531353473663,
0.07106044143438339,
-0.01044827327132225,
-0.06977739930152893,
0.11272217333316803,
0.015659291297197342,
-0.03067840449512005,
0.021955255419015884,
0.06461543589830399,
-0.0716981291770935,
-0.02817065455019474,
-0.06754184514284134,
0.13719263672828674,
-0.08612000197172165,
-0.03902946040034294,
-0.08327492326498032,
-0.07082365453243256,
-0.03257657214999199,
0.09914835542440414,
0.062177058309316635,
-0.011092182248830795,
-0.014385067857801914,
-0.02562425471842289,
-0.10685867816209793,
0.08523236215114594,
0.039405278861522675,
0.07814429700374603,
-0.20603284239768982,
-0.030703669413924217,
0.05563230812549591,
-0.028979875147342682,
-0.06726782768964767,
-0.021924879401922226,
-0.07509555667638779,
-0.02428257465362549,
-0.04582654684782028,
0.10391157865524292,
-0.09901858866214752,
0.007180153392255306,
-0.04038698226213455,
-0.03781870752573013,
-0.015500243753194809,
0.04643344134092331,
-0.013886834494769573,
-0.021665258333086967,
0.005997329019010067,
0.04266213998198509,
-0.10093989968299866,
-0.06927894055843353,
-0.02179509401321411,
-0.07151764631271362,
0.023306097835302353,
-0.030765434727072716,
-0.03506799042224884,
-0.00142877537291497,
-0.19826927781105042,
0.05676979944109917,
0.08950874209403992,
-0.001832907204516232,
-0.005028935614973307,
-0.011682429350912571,
0.015350023284554482,
-0.030704835429787636,
0.005948605947196484,
-0.02725050039589405,
-0.011469719931483269,
-0.10233753174543381,
0.02998560480773449,
-0.020079730078577995,
0.02709435671567917,
-0.051088761538267136,
0.06004020944237709,
0.15954020619392395,
0.09050803631544113,
0.09843651950359344,
-0.08764635026454926,
0.06654002517461777,
-0.1621038168668747,
-0.011386245489120483,
-0.003017950337380171,
-0.04593811556696892,
-0.03079632855951786,
0.01790465973317623,
0.015510322526097298,
-0.006736717186868191,
0.10814527422189713,
0.029536938294768333,
-0.03573894500732422,
-0.020247936248779297,
-0.001094343955628574,
0.07219237834215164,
-0.01292289886623621,
0.1925036609172821,
0.024291055276989937,
0.026994992047548294,
-0.029334543272852898,
0.06168104335665703,
0.0977795273065567,
0.058701809495687485,
-0.003029116429388523,
0.08011338114738464,
-0.01428475696593523,
0.0850086510181427,
0.023578263819217682,
-0.00036789727164432406,
-0.04833542928099632,
0.09629721939563751,
-0.04832508787512779,
0.046230778098106384,
-0.043133534491062164,
0.040466733276844025,
0.1016550064086914,
-0.1255420446395874,
-0.009398452006280422,
0.13597574830055237,
-0.037953875958919525,
-0.11323592066764832,
-0.2071547508239746,
-0.08205140382051468,
-0.15772491693496704,
0.021330315619707108,
-0.07585472613573074,
0.003336096415296197,
0.06799867749214172,
-0.024553662165999413,
0.04525020718574524,
0.15400540828704834,
-0.014432835392653942,
-0.009553452022373676,
0.050849899649620056,
-0.015235833823680878,
-0.03499477729201317,
0.02939322218298912,
-0.03333459794521332,
0.09604672342538834,
-0.017109068110585213,
0.006435131188482046,
0.030694741755723953,
0.04049261659383774,
0.07751652598381042,
0.020736973732709885,
-0.08531073480844498,
-0.023872654885053635,
0.032121580094099045,
-0.004823280498385429,
0.14755402505397797,
0.06826546788215637,
-0.027638951316475868,
-0.04505983367562294,
0.14115653932094574,
-0.050764039158821106,
0.005861143581569195,
-0.13634136319160461,
0.08400274813175201,
-0.05307725816965103,
0.014357130974531174,
-0.007088846527040005,
-0.09816235303878784,
-0.0171080082654953,
0.1311597228050232,
0.14342835545539856,
0.013697715476155281,
0.020365944132208824,
-0.0704905316233635,
-0.023051613941788673,
-0.05906729772686958,
0.050136204808950424,
0.0037438306026160717,
0.16511593759059906,
-0.050339438021183014,
0.06569021195173264,
-0.04198587313294411,
-0.06544005125761032,
-0.03932235389947891,
-0.018128590658307076,
-0.07646144181489944,
-0.015382393263280392,
-0.02594980224967003,
0.056456007063388824,
-0.08493974059820175,
-0.10089974105358124,
0.16640225052833557,
-0.09472783654928207,
-0.034177035093307495,
-0.0834641233086586,
0.009133431129157543,
0.05081462115049362,
0.009616252966225147,
-0.06570569425821304,
-0.0022899082396179438,
0.04382942616939545,
-0.037988826632499695,
-0.11843518912792206,
-0.059298623353242874,
-0.026765691116452217,
-0.11207162588834763,
0.20796455442905426,
-0.03856281563639641,
-0.03308695927262306,
0.0449577197432518,
-0.009171932935714722,
-0.12186035513877869,
0.08106590807437897,
-0.009530269540846348,
-0.11255717277526855,
-0.026332374662160873,
0.12610416114330292,
-0.08030693978071213,
0.16938374936580658,
0.010877957567572594,
-0.11020255088806152,
-0.00006721775571350008,
0.036391884088516235,
-0.07054207473993301,
-0.10131444036960602,
-0.012952281162142754,
-0.09402424842119217,
0.13324657082557678,
0.092694953083992,
-0.01703512854874134,
-0.01993323117494583,
-0.012222444638609886,
0.06406253576278687,
0.04021606594324112,
0.04778039827942848,
0.06201034039258957,
-0.07572969794273376,
-0.0328904390335083,
0.0444820374250412,
0.03440770506858826,
-0.2674715220928192,
-0.016742568463087082,
-0.14984022080898285,
-0.04856058582663536,
0.006771868094801903,
0.04117171838879585,
0.15939836204051971,
0.04244844242930412,
-0.018410349264740944,
-0.20317667722702026,
0.01118440367281437,
0.11258535832166672,
-0.11954115331172943,
-0.0752624124288559
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | GowthamMl/deepseeker-finetuned-v1 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:52:49+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# nslPOS
This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 2
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.36.0
- Pytorch 2.1.2+cu118
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "base_model": "naver-clova-ix/donut-base", "model-index": [{"name": "nslPOS", "results": []}]} | null | saniasinghania/nslPOS | [
"transformers",
"tensorboard",
"safetensors",
"vision-encoder-decoder",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:naver-clova-ix/donut-base",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:54:47+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #dataset-imagefolder #base_model-naver-clova-ix/donut-base #license-mit #endpoints_compatible #region-us
|
# nslPOS
This model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 2
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.36.0
- Pytorch 2.1.2+cu118
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# nslPOS\n\nThis model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.36.0\n- Pytorch 2.1.2+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #dataset-imagefolder #base_model-naver-clova-ix/donut-base #license-mit #endpoints_compatible #region-us \n",
"# nslPOS\n\nThis model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.36.0\n- Pytorch 2.1.2+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
70,
34,
6,
12,
8,
3,
103,
4,
35
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #dataset-imagefolder #base_model-naver-clova-ix/donut-base #license-mit #endpoints_compatible #region-us \n# nslPOS\n\nThis model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.36.0\n- Pytorch 2.1.2+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.06370986253023148,
0.1284140646457672,
-0.0021314325276762247,
0.06557737290859222,
0.13262328505516052,
0.013606552965939045,
0.10250620543956757,
0.10652536153793335,
-0.097194142639637,
0.08107033371925354,
0.05831793323159218,
0.033611416816711426,
0.056380484253168106,
0.16181710362434387,
-0.03350641205906868,
-0.25326359272003174,
0.041804712265729904,
0.0073025524616241455,
-0.02870056964457035,
0.08850415050983429,
0.07399427890777588,
-0.10537586361169815,
0.05637795478105545,
-0.007817772217094898,
-0.16157515347003937,
0.020734550431370735,
-0.05015541613101959,
-0.059611666947603226,
0.06750886887311935,
-0.0034318596590310335,
0.12298784404993057,
0.029632657766342163,
0.14228177070617676,
-0.23385165631771088,
0.003512305673211813,
0.09502077847719193,
0.02843518927693367,
0.06004795432090759,
0.0691695585846901,
0.01871475763618946,
0.09208549559116364,
-0.18512792885303497,
0.11389800161123276,
0.014139524661004543,
-0.0675472617149353,
-0.15154516696929932,
-0.08133738487958908,
0.07996241003274918,
0.11837299913167953,
0.10304374992847443,
0.004679015837609768,
0.1422959715127945,
-0.07015203684568405,
0.07382619380950928,
0.16635000705718994,
-0.22407187521457672,
-0.07394788414239883,
0.07774055749177933,
0.03635922446846962,
0.07303375005722046,
-0.10626161843538284,
0.0012795819202437997,
0.05867042392492294,
0.029456784948706627,
0.0770556852221489,
-0.0021023184526711702,
-0.062090493738651276,
-0.03397554159164429,
-0.14965026080608368,
-0.03267623856663704,
0.12684814631938934,
0.06964770704507828,
-0.037706974893808365,
-0.09411536157131195,
-0.04940390586853027,
-0.10206048935651779,
-0.030322181060910225,
-0.060084812343120575,
0.04672011733055115,
-0.0392020083963871,
-0.060593381524086,
-0.06616313755512238,
-0.08534789085388184,
-0.04247294366359711,
-0.00018345213902648538,
0.07236403226852417,
0.05000906065106392,
0.010872527956962585,
-0.019263969734311104,
0.10247565805912018,
-0.019118232652544975,
-0.10459823161363602,
-0.020215924829244614,
-0.003385401563718915,
-0.08635248243808746,
-0.07569479942321777,
-0.013241144828498363,
-0.04384956881403923,
-0.008317136205732822,
0.11192143708467484,
-0.07233122736215591,
0.08208050578832626,
-0.04134926572442055,
0.013459722511470318,
-0.028236018493771553,
0.11621107161045074,
-0.018752852454781532,
0.010718366131186485,
-0.007744538132101297,
0.0772053673863411,
0.018520891666412354,
-0.01057012565433979,
-0.0850144550204277,
-0.005607987754046917,
0.09182875603437424,
0.0548536479473114,
-0.04908289760351181,
0.029026487842202187,
-0.01817115768790245,
-0.020687729120254517,
0.002254818333312869,
-0.1182832345366478,
0.07270263135433197,
-0.014445827342569828,
-0.05740541219711304,
0.020570138469338417,
0.04716324433684349,
0.008586244657635689,
-0.029252948239445686,
0.09730420261621475,
-0.052533719688653946,
0.018538016825914383,
-0.09969929605722427,
-0.052837636321783066,
0.04426151514053345,
-0.06439348310232162,
-0.026309002190828323,
-0.07954911887645721,
-0.1829979419708252,
-0.04424409940838814,
0.04376952350139618,
-0.04204993695020676,
-0.018831945955753326,
-0.05981215834617615,
-0.06757200509309769,
0.004497208166867495,
-0.00010588121222099289,
0.09702300280332565,
-0.0493810772895813,
0.05722491443157196,
-0.01525349821895361,
0.029303932562470436,
0.017549922689795494,
0.025022128596901894,
-0.08397955447435379,
0.03508240357041359,
-0.15826085209846497,
0.06589394062757492,
-0.08613377809524536,
0.029144784435629845,
-0.10858754813671112,
-0.10856414586305618,
-0.05097416043281555,
-0.02900267392396927,
0.033732496201992035,
0.10326942056417465,
-0.1858842372894287,
-0.00749632716178894,
0.14484542608261108,
-0.10579950362443924,
-0.04887121170759201,
0.10735895484685898,
-0.043611202389001846,
-0.001556043396703899,
0.05629021301865578,
0.16205769777297974,
0.07855939120054245,
-0.15141217410564423,
-0.005611345637589693,
-0.041495539247989655,
0.04184774309396744,
0.009437307715415955,
0.03861817717552185,
-0.006979080848395824,
0.038428544998168945,
-0.001806630869396031,
-0.0600237175822258,
0.00540607376024127,
-0.08285340666770935,
-0.08041848987340927,
-0.06140730902552605,
-0.057876504957675934,
0.005570500623434782,
0.03524048626422882,
0.027703847736120224,
-0.086142897605896,
-0.11400048434734344,
0.11028986424207687,
0.10766872018575668,
-0.07513374835252762,
0.02512407675385475,
-0.06991912424564362,
0.020982326939702034,
-0.05866297334432602,
-0.035925041884183884,
-0.1620892882347107,
-0.14626619219779968,
0.025925248861312866,
-0.09699922055006027,
0.04085855558514595,
-0.01269660796970129,
0.05368141829967499,
0.08445560932159424,
-0.03986307233572006,
-0.03003278188407421,
-0.08293802291154861,
0.024428730830550194,
-0.07531580328941345,
-0.17671461403369904,
-0.04165045917034149,
-0.021244151517748833,
0.14592452347278595,
-0.2576034665107727,
0.022445999085903168,
0.007038692943751812,
0.1716315746307373,
0.029816266149282455,
-0.059676267206668854,
-0.005220227409154177,
0.06236787512898445,
0.0036854189820587635,
-0.1067667305469513,
0.036806683987379074,
0.000628830399364233,
-0.058405809104442596,
-0.07125350832939148,
-0.16229133307933807,
0.0070985956117510796,
0.09577881544828415,
0.03981054946780205,
-0.09539896994829178,
-0.016139233484864235,
-0.06560689210891724,
-0.04360378161072731,
-0.10444580018520355,
0.02555275708436966,
0.1397731602191925,
0.0003979077737312764,
0.1049354076385498,
-0.046368308365345,
-0.049653079360723495,
0.01756487414240837,
-0.0009942473843693733,
-0.05634754151105881,
0.059866923838853836,
0.09275133907794952,
-0.12295099347829819,
0.10325855016708374,
0.10012950003147125,
-0.04630442336201668,
0.13817180693149567,
-0.04686018452048302,
-0.08873231709003448,
-0.013288983143866062,
0.03554270416498184,
-0.0019012584816664457,
0.1277695596218109,
-0.06380945444107056,
0.015984592959284782,
0.015205408446490765,
0.02027788758277893,
0.03831332176923752,
-0.18870602548122406,
0.006958636920899153,
0.03184938803315163,
-0.04021824896335602,
0.037382110953330994,
-0.0380437970161438,
0.03817009925842285,
0.07649954408407211,
0.008851229213178158,
-0.021320626139640808,
0.026062067598104477,
-0.0016124972607940435,
-0.08074096590280533,
0.1501549333333969,
-0.1064942479133606,
-0.14527255296707153,
-0.11544298380613327,
0.05606096610426903,
-0.041095882654190063,
-0.02829396352171898,
0.028417624533176422,
-0.09073372930288315,
-0.06857898831367493,
-0.11823198944330215,
-0.04265889152884483,
-0.039034388959407806,
-0.01998434029519558,
0.07102804630994797,
0.01857246272265911,
0.06808657199144363,
-0.10341282933950424,
-0.006227843463420868,
-0.01472317986190319,
-0.08255233615636826,
0.013349536806344986,
0.044687606394290924,
0.11450246721506119,
0.1330834925174713,
-0.04180910065770149,
0.02683519758284092,
-0.033143289387226105,
0.16966408491134644,
-0.08038938045501709,
0.013478090986609459,
0.09644481539726257,
0.008519627153873444,
0.056269992142915726,
0.12461665272712708,
0.023860331624746323,
-0.09390699863433838,
0.03197255730628967,
0.08177292346954346,
-0.03026207536458969,
-0.2025837004184723,
-0.05850067734718323,
-0.03092402033507824,
-0.0754866972565651,
0.09547898173332214,
0.0589238777756691,
0.03926421329379082,
0.0404793843626976,
-0.0230746790766716,
0.026879968121647835,
0.005178677849471569,
0.0999683290719986,
0.11566490679979324,
0.04301398992538452,
0.07773764431476593,
-0.04261774942278862,
-0.04770980030298233,
0.05310342088341713,
0.008582445792853832,
0.26326364278793335,
0.00486579118296504,
0.06600277125835419,
0.03673327714204788,
0.1326752007007599,
0.0018501811427995563,
-0.001195744494907558,
0.04516606405377388,
-0.007828335277736187,
0.02070130780339241,
-0.06928232312202454,
0.006786673329770565,
0.027731547132134438,
-0.021167762577533722,
0.0321369469165802,
-0.0839948058128357,
0.04032451659440994,
0.026286548003554344,
0.2364174723625183,
0.029284629970788956,
-0.31914466619491577,
-0.051101069897413254,
0.018406299874186516,
-0.010156880132853985,
-0.05642769858241081,
0.0004957097116857767,
0.1232246682047844,
-0.11975909769535065,
0.08094896376132965,
-0.06571445614099503,
0.08428891748189926,
-0.04282538965344429,
-0.00543360598385334,
0.05590696260333061,
0.11836103349924088,
-0.0032816568855196238,
0.07597946375608444,
-0.1867956966161728,
0.20188575983047485,
0.006920794956386089,
0.1319357007741928,
-0.04903476685285568,
0.029395505785942078,
0.014735647477209568,
0.09600341320037842,
0.12195291370153427,
-0.008459948003292084,
-0.138351172208786,
-0.17897358536720276,
-0.09334030002355576,
0.02518323063850403,
0.12455902248620987,
0.003011091146618128,
0.0824689120054245,
-0.022410720586776733,
-0.0029604043811559677,
0.05042768642306328,
-0.04271111264824867,
-0.22021281719207764,
-0.12940719723701477,
-0.0010431332048028708,
0.036934446543455124,
-0.021905897185206413,
-0.08168906718492508,
-0.1106446385383606,
-0.05763552710413933,
0.18080885708332062,
0.027754317969083786,
-0.02173311822116375,
-0.14248257875442505,
0.08144185692071915,
0.09399309009313583,
-0.03949042409658432,
0.019546138122677803,
0.024410905316472054,
0.15311945974826813,
0.018763577565550804,
-0.09282106906175613,
0.05815384164452553,
-0.06340888887643814,
-0.13885429501533508,
-0.05465969443321228,
0.12036813795566559,
0.05203565955162048,
0.03511229529976845,
0.0031361111905425787,
0.03279787302017212,
0.058906637132167816,
-0.07959383726119995,
0.02325999177992344,
0.10728085041046143,
0.08202540874481201,
0.05701929330825806,
-0.07713691145181656,
-0.013305178843438625,
-0.03959584981203079,
-0.01341365184634924,
0.11780636757612228,
0.18623779714107513,
-0.0717739462852478,
0.05739492550492287,
0.03805778920650482,
-0.09139293432235718,
-0.15894456207752228,
0.10433026403188705,
0.08022692054510117,
-0.0028427408542484045,
0.030866870656609535,
-0.14411844313144684,
0.11116564273834229,
0.14412790536880493,
-0.02076694369316101,
0.04734010621905327,
-0.33046647906303406,
-0.12309309095144272,
0.07402217388153076,
0.14102932810783386,
-0.021027186885476112,
-0.13949884474277496,
-0.02522735856473446,
-0.01952355168759823,
-0.15391701459884644,
0.13774177432060242,
-0.08711805194616318,
0.08953046798706055,
0.010909355245530605,
0.06675953418016434,
0.01743370108306408,
-0.03988125920295715,
0.13734155893325806,
0.009687921032309532,
0.09149286895990372,
-0.07663944363594055,
0.06170567125082016,
0.07581166177988052,
-0.0841655358672142,
0.05866583436727524,
0.014197438023984432,
0.034704238176345825,
-0.10640982538461685,
-0.02482185885310173,
-0.05153956264257431,
0.06286279112100601,
-0.04991545528173447,
-0.05071066692471504,
-0.047896839678287506,
0.06357324868440628,
0.07426176220178604,
-0.033160753548145294,
0.09459513425827026,
0.022156864404678345,
0.10567821562290192,
0.0744343027472496,
0.07468938827514648,
-0.009627039544284344,
-0.09519464522600174,
-0.010520607233047485,
-0.03074675239622593,
0.05879475176334381,
-0.09446113556623459,
0.019420906901359558,
0.12934531271457672,
0.024330787360668182,
0.14244334399700165,
0.03956269100308418,
-0.061000674962997437,
-0.003607987193390727,
0.03540428727865219,
-0.10244293510913849,
-0.17715753614902496,
0.010598783381283283,
-0.02356533333659172,
-0.0811217874288559,
0.0067303357645869255,
0.08760082721710205,
-0.08873342722654343,
0.006856490857899189,
-0.01499198842793703,
0.03161390498280525,
-0.016388079151511192,
0.15313464403152466,
0.027280723676085472,
0.048480045050382614,
-0.06804250180721283,
0.13334587216377258,
0.0894973874092102,
-0.13267894089221954,
0.057628605514764786,
0.09132292866706848,
-0.07785356789827347,
-0.02371075190603733,
0.11869601160287857,
0.19218270480632782,
-0.016676461324095726,
-0.06310052424669266,
-0.06766200810670853,
-0.10585955530405045,
0.04812393710017204,
0.10849176347255707,
0.03917131572961807,
-0.029007799923419952,
-0.01370901707559824,
0.037290703505277634,
-0.16589359939098358,
0.07758861035108566,
0.036624450236558914,
0.09154222160577774,
-0.16439683735370636,
0.10124809294939041,
0.017430897802114487,
0.021849019452929497,
-0.01823103055357933,
0.011958763003349304,
-0.08368027210235596,
-0.010242613963782787,
-0.07939998805522919,
0.011623288504779339,
-0.047307830303907394,
0.0023489256855100393,
-0.0038858286570757627,
-0.027186725288629532,
-0.04360981285572052,
0.055307332426309586,
-0.0573192797601223,
-0.06438802182674408,
0.014082871377468109,
0.053824152797460556,
-0.12946943938732147,
-0.006510620005428791,
0.004795253276824951,
-0.08805977553129196,
0.028226910158991814,
0.03128657117486,
0.02001921460032463,
0.017548417672514915,
-0.16495811939239502,
-0.018151726573705673,
0.07071053981781006,
0.017843756824731827,
0.048511240631341934,
-0.08343330770730972,
-0.01821969449520111,
0.008873303420841694,
0.03751430660486221,
0.014990510419011116,
0.055957481265068054,
-0.11540684103965759,
-0.044985879212617874,
-0.07518994063138962,
-0.058410391211509705,
-0.04221618175506592,
0.045407891273498535,
0.10302802175283432,
0.02217561937868595,
0.15814518928527832,
-0.10559292882680893,
0.04732099920511246,
-0.18799668550491333,
-0.0476316474378109,
0.0028090502601116896,
-0.014524774625897408,
-0.01231760997325182,
-0.01767430454492569,
0.069815993309021,
-0.05303670093417168,
0.09529101848602295,
0.0008507549646310508,
0.08836967498064041,
0.05028129369020462,
-0.06577478349208832,
-0.038159873336553574,
0.03238960728049278,
0.18671028316020966,
0.052221380174160004,
-0.031054755672812462,
0.10711003839969635,
-0.019658327102661133,
0.08523420244455338,
0.06150420010089874,
0.18797709047794342,
0.1840381622314453,
-0.040885526686906815,
0.07294712215662003,
0.07755865156650543,
-0.07539819180965424,
-0.13986702263355255,
0.13341039419174194,
-0.02946535497903824,
0.12392135709524155,
-0.050294142216444016,
0.11555521935224533,
0.10622869431972504,
-0.1798797994852066,
0.028706533834338188,
-0.039387136697769165,
-0.1016097217798233,
-0.09562458097934723,
-0.11130189895629883,
-0.08944670855998993,
-0.12597566843032837,
0.03664764389395714,
-0.10287080705165863,
0.02772785723209381,
0.09204297512769699,
0.027544554322957993,
0.014429298229515553,
0.17516258358955383,
-0.04055779427289963,
0.025285277515649796,
0.05738532170653343,
0.009210839867591858,
-0.01443096250295639,
-0.07087258249521255,
-0.06286296993494034,
0.06223819777369499,
0.0016254318179562688,
0.07629688829183578,
-0.029728243127465248,
0.02137698233127594,
0.04934391751885414,
-0.012080547399818897,
-0.07046350091695786,
0.03766971826553345,
0.02134404517710209,
0.03877279534935951,
0.029021883383393288,
0.05219881981611252,
-0.023421883583068848,
-0.03836890310049057,
0.2443612813949585,
-0.05850525572896004,
-0.09876435250043869,
-0.11467042565345764,
0.1549101620912552,
0.02952473796904087,
-0.011223883368074894,
0.05548970028758049,
-0.12544772028923035,
0.007081573363393545,
0.11632100492715836,
0.12290287017822266,
-0.015294447541236877,
-0.00995013490319252,
-0.019594738259911537,
-0.020105144008994102,
-0.053672268986701965,
0.11129017919301987,
0.10423493385314941,
0.01568855531513691,
-0.05026259273290634,
-0.0158204585313797,
-0.012644981034100056,
-0.04375597462058067,
-0.10849907249212265,
0.03352378308773041,
0.021135421469807625,
0.0005761846550740302,
-0.0426563061773777,
0.10009318590164185,
0.0074563766829669476,
-0.12923292815685272,
0.09357297420501709,
-0.15208286046981812,
-0.15521003305912018,
-0.002707055304199457,
0.13159748911857605,
-0.0451476015150547,
0.02255346067249775,
-0.03248513862490654,
-0.017571916803717613,
0.10659503191709518,
-0.009416336193680763,
-0.05109713599085808,
-0.09538086503744125,
0.02540878765285015,
-0.08309008181095123,
0.2520201802253723,
0.010716372169554234,
0.07174893468618393,
0.09033568948507309,
0.022547677159309387,
-0.12807965278625488,
0.06294899433851242,
0.04905954375863075,
-0.05137409642338753,
0.024280326440930367,
0.1712561845779419,
-0.06456681340932846,
0.09221092611551285,
0.05388522893190384,
-0.1686372011899948,
-0.010732659138739109,
-0.0550127811729908,
-0.014569258317351341,
-0.059452567249536514,
0.0019044667715206742,
-0.08207330852746964,
0.14568860828876495,
0.1687348634004593,
-0.013739001005887985,
0.022212078794836998,
-0.06966076791286469,
0.03632538393139839,
0.04722565412521362,
0.1061856597661972,
-0.004317686893045902,
-0.18476559221744537,
0.019297920167446136,
-0.03356616944074631,
0.0348815955221653,
-0.22153665125370026,
-0.12218721956014633,
0.012193264439702034,
-0.057910360395908356,
-0.07269943505525589,
0.10867705196142197,
0.10732759535312653,
0.03131716698408127,
-0.05030577629804611,
-0.15803417563438416,
-0.02430657669901848,
0.15139704942703247,
-0.13103161752223969,
-0.030804231762886047
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-large-cased-lora-1.58M-snli-model2
This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8331
- Accuracy: 0.687
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 256
- eval_batch_size: 256
- seed: 6
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.5164 | 1.0 | 2146 | 0.4262 | 0.8406 |
| 0.4687 | 2.0 | 4292 | 0.3904 | 0.8540 |
| 0.4562 | 3.0 | 6438 | 0.3824 | 0.8575 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-large-cased", "model-index": [{"name": "bert-large-cased-lora-1.58M-snli-model2", "results": []}]} | text-classification | varun-v-rao/bert-large-cased-lora-1.58M-snli-model2 | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:bert-large-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T06:57:08+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-large-cased-lora-1.58M-snli-model2
=======================================
This model is a fine-tuned version of bert-large-cased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8331
* Accuracy: 0.687
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 256
* eval\_batch\_size: 256
* seed: 6
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 6\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 6\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
68,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 6\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.08380687236785889,
0.09604667127132416,
-0.0022442650515586138,
0.1115875169634819,
0.14797595143318176,
0.022443508729338646,
0.15785902738571167,
0.10879659652709961,
-0.06997837126255035,
0.050637874752283096,
0.12808381021022797,
0.12629282474517822,
0.012304246425628662,
0.12039896100759506,
-0.06541161984205246,
-0.22940978407859802,
0.00501704216003418,
0.0354124940931797,
-0.07027395069599152,
0.11373141407966614,
0.09139653295278549,
-0.13154670596122742,
0.09555875509977341,
-0.008379251696169376,
-0.18652237951755524,
0.01450068037956953,
0.019215870648622513,
-0.05388394743204117,
0.13300904631614685,
0.037687286734580994,
0.13750894367694855,
0.01825219951570034,
0.09317469596862793,
-0.2010529339313507,
0.008286776021122932,
0.057660218328237534,
-0.00205851998180151,
0.08284591138362885,
0.037596601992845535,
0.01960039883852005,
0.08278557658195496,
-0.08890887349843979,
0.062491461634635925,
0.01614416390657425,
-0.11445307731628418,
-0.1963450014591217,
-0.07084780186414719,
0.0399734228849411,
0.08578652888536453,
0.07925687730312347,
-0.009791651740670204,
0.12081831693649292,
-0.052837301045656204,
0.08757985383272171,
0.21846599876880646,
-0.3193298280239105,
-0.06559756398200989,
0.056147437542676926,
0.03372372314333916,
0.0855480507016182,
-0.10435139387845993,
-0.018771149218082428,
0.07431026548147202,
0.026250043883919716,
0.135457381606102,
-0.02876865863800049,
-0.07086043059825897,
0.010261737741529942,
-0.14986324310302734,
-0.020402198657393456,
0.14984571933746338,
0.04825163260102272,
-0.041091132909059525,
-0.052068959921598434,
-0.06384028494358063,
-0.1502360999584198,
-0.03683549910783768,
-0.025654921308159828,
0.05538029223680496,
-0.027482669800519943,
-0.06373361498117447,
-0.012339182198047638,
-0.10842719674110413,
-0.08097540587186813,
-0.06592027842998505,
0.1430017501115799,
0.03856480121612549,
0.0018338392255827785,
-0.018717437982559204,
0.09379129111766815,
-0.03504263982176781,
-0.12105691432952881,
0.021665219217538834,
0.02618010900914669,
0.010021145455539227,
-0.05784078314900398,
-0.054840218275785446,
-0.07199633866548538,
0.025035029277205467,
0.1304887980222702,
-0.053321316838264465,
0.04987237602472305,
0.013632054440677166,
0.04814700037240982,
-0.09860428422689438,
0.17509302496910095,
-0.0410127229988575,
-0.018810143694281578,
0.009035496972501278,
0.0751812607049942,
0.04108782485127449,
-0.0031774311792105436,
-0.13190394639968872,
0.023314103484153748,
0.10324938595294952,
0.013336929492652416,
-0.08245635777711868,
0.07984985411167145,
-0.04689313843846321,
-0.0011390929576009512,
0.007887410931289196,
-0.0864446759223938,
0.03186626359820366,
0.00336472038179636,
-0.0514487624168396,
-0.0585523284971714,
0.023416584357619286,
0.012344678863883018,
0.005874453578144312,
0.11459033936262131,
-0.09089917689561844,
0.008632726036012173,
-0.09016494452953339,
-0.12357147783041,
0.02504757046699524,
-0.08337150514125824,
0.022816024720668793,
-0.11205493658781052,
-0.15460671484470367,
-0.00210530124604702,
0.052882518619298935,
-0.02715977653861046,
-0.03157297894358635,
-0.051991384476423264,
-0.0724773108959198,
0.015219991095364094,
-0.017679978162050247,
0.08813786506652832,
-0.0637628436088562,
0.09123452007770538,
0.04050429165363312,
0.06988662481307983,
-0.05222303047776222,
0.035923704504966736,
-0.09537896513938904,
0.028408292680978775,
-0.17272686958312988,
0.014391805976629257,
-0.06446699053049088,
0.05766747519373894,
-0.08466947823762894,
-0.07875711470842361,
-0.00681255292147398,
0.01474655419588089,
0.06887245923280716,
0.08383956551551819,
-0.16398701071739197,
-0.06817081570625305,
0.17946843802928925,
-0.08782880753278732,
-0.14210110902786255,
0.13108479976654053,
-0.05658743157982826,
0.04965770244598389,
0.06343235075473785,
0.17641036212444305,
0.05486083775758743,
-0.09457173943519592,
-0.008830961771309376,
0.003574403701350093,
0.0608506053686142,
-0.049815401434898376,
0.06671011447906494,
0.003648392390459776,
-0.006990899331867695,
0.01583258993923664,
-0.044353023171424866,
0.04855528473854065,
-0.07952975481748581,
-0.08360765874385834,
-0.040134407579898834,
-0.0994732454419136,
0.047251127660274506,
0.05371067672967911,
0.06711933761835098,
-0.11217445135116577,
-0.08737211674451828,
0.0706159844994545,
0.06793440133333206,
-0.07492745667695999,
0.017948640510439873,
-0.07121937721967697,
0.07885425537824631,
-0.06643889844417572,
-0.014760901220142841,
-0.15092183649539948,
-0.050243496894836426,
0.020680131390690804,
-0.008837911300361156,
0.018425652757287025,
0.010586042888462543,
0.07106166332960129,
0.07602950185537338,
-0.06980995088815689,
-0.01902131363749504,
-0.01791686937212944,
0.015187456272542477,
-0.13066518306732178,
-0.19910664856433868,
-0.018414827063679695,
-0.03086831234395504,
0.12299209088087082,
-0.23748555779457092,
0.04962348937988281,
0.00323134521022439,
0.07871189713478088,
0.0352950245141983,
-0.0029553910717368126,
-0.05095444992184639,
0.07365338504314423,
-0.04494110494852066,
-0.0533142015337944,
0.05709881708025932,
0.005751322489231825,
-0.08834981918334961,
-0.041774313896894455,
-0.14519338309764862,
0.18886961042881012,
0.13395051658153534,
-0.09268999844789505,
-0.08071331679821014,
-0.012703928165137768,
-0.043156497180461884,
-0.02835635468363762,
-0.05311909690499306,
-0.0007479380001313984,
0.1385187804698944,
-0.019835207611322403,
0.15174223482608795,
-0.0853712186217308,
-0.03686077147722244,
0.028221391141414642,
-0.04582173004746437,
0.0070360018871724606,
0.11001938581466675,
0.12261948734521866,
-0.11064068228006363,
0.1523081213235855,
0.18171334266662598,
-0.10525399446487427,
0.1392822265625,
-0.043203581124544144,
-0.06096820533275604,
-0.021285751834511757,
0.0027065561152994633,
0.008176037110388279,
0.11227069050073624,
-0.12943808734416962,
-0.0030058720149099827,
0.005930486600846052,
0.014761489816009998,
0.015619810670614243,
-0.22014115750789642,
-0.025045769289135933,
0.039155423641204834,
-0.04899090155959129,
0.012116264551877975,
-0.024381423369050026,
-0.008131368085741997,
0.10155372321605682,
-0.000955894822254777,
-0.08904344588518143,
0.043408408761024475,
-0.004670440219342709,
-0.08565884083509445,
0.2130289077758789,
-0.0750863254070282,
-0.13756892085075378,
-0.13784196972846985,
-0.06964075565338135,
-0.03783957287669182,
0.02360568195581436,
0.060867585241794586,
-0.08147396147251129,
-0.039846304804086685,
-0.10676544904708862,
0.011988264508545399,
0.031556159257888794,
0.038107678294181824,
0.018605316057801247,
0.005156445316970348,
0.07954005151987076,
-0.10277274250984192,
-0.012447327375411987,
-0.05090980604290962,
-0.06549505144357681,
0.02663971669971943,
0.027831420302391052,
0.11893121898174286,
0.14727507531642914,
-0.032944582402706146,
-0.002347763627767563,
-0.035795845091342926,
0.21301570534706116,
-0.05622468888759613,
-0.023203492164611816,
0.12034258991479874,
-0.030823267996311188,
0.04425952211022377,
0.1368197500705719,
0.06854206323623657,
-0.09134816378355026,
0.020262842997908592,
0.041627999395132065,
-0.03128517046570778,
-0.21862368285655975,
-0.0374707356095314,
-0.03776581585407257,
0.017930366098880768,
0.1020926907658577,
0.03477520868182182,
0.023652411997318268,
0.06680536270141602,
0.03760275989770889,
0.08782066404819489,
-0.01892249472439289,
0.07065995037555695,
0.12128318101167679,
0.042829565703868866,
0.12629102170467377,
-0.04953876882791519,
-0.06717834621667862,
0.03505203127861023,
0.005277441348880529,
0.20864327251911163,
0.029139695689082146,
0.14033499360084534,
0.05659429356455803,
0.1444542557001114,
0.0018030463252216578,
0.07179144769906998,
-0.012975004501640797,
-0.04664574936032295,
-0.014818716794252396,
-0.050420861691236496,
-0.02149115316569805,
0.04143840819597244,
-0.08734458684921265,
0.05806916579604149,
-0.11366492509841919,
0.02003439888358116,
0.059401996433734894,
0.23440156877040863,
0.04579682648181915,
-0.324935644865036,
-0.08393487334251404,
0.02434929832816124,
-0.03130302205681801,
-0.018060456961393356,
0.038376584649086,
0.14218492805957794,
-0.051138147711753845,
0.024755161255598068,
-0.07155497372150421,
0.07835608720779419,
-0.04647327959537506,
0.0472068153321743,
0.07042475789785385,
0.09061966091394424,
-0.006131492089480162,
0.07183248549699783,
-0.27268049120903015,
0.27643680572509766,
0.012315810658037663,
0.06683941930532455,
-0.05625626817345619,
0.003431141609326005,
0.029962072148919106,
0.07943984121084213,
0.0762842521071434,
-0.02392670325934887,
-0.06177129969000816,
-0.2004920244216919,
-0.05327153205871582,
0.02937711402773857,
0.09961120784282684,
-0.031819477677345276,
0.09928008168935776,
-0.034763578325510025,
0.0022487277165055275,
0.09181012213230133,
-0.008370666764676571,
-0.07193461805582047,
-0.10023599863052368,
-0.01962127536535263,
0.04529890790581703,
-0.043343912810087204,
-0.07976215332746506,
-0.10498712211847305,
-0.1373256891965866,
0.15006111562252045,
-0.05790797993540764,
-0.020008845254778862,
-0.09098351746797562,
0.0703556165099144,
0.04886096715927124,
-0.07803850620985031,
0.050204310566186905,
0.00910107884556055,
0.09139300882816315,
0.02153187245130539,
-0.058260172605514526,
0.12936051189899445,
-0.07549040019512177,
-0.17051449418067932,
-0.07684173434972763,
0.09767340123653412,
0.018078554421663284,
0.04734295234084129,
-0.002402189653366804,
0.0031714988872408867,
-0.006348428316414356,
-0.0811537578701973,
0.02354496531188488,
-0.0028217320796102285,
0.062360234558582306,
0.012320290319621563,
-0.07913350313901901,
-0.0068260179832577705,
-0.04643557593226433,
-0.03131168335676193,
0.1689814180135727,
0.2655394673347473,
-0.09300430864095688,
-0.00553617998957634,
0.059733256697654724,
-0.07263793796300888,
-0.21305188536643982,
0.0437292642891407,
0.035151638090610504,
0.0038321721367537975,
0.04130544885993004,
-0.14704637229442596,
0.12829910218715668,
0.10617391765117645,
-0.02344527095556259,
0.10117131471633911,
-0.29060593247413635,
-0.13359516859054565,
0.1302434355020523,
0.15624433755874634,
0.10754630714654922,
-0.15429896116256714,
-0.03192361444234848,
-0.035353884100914,
-0.1055215522646904,
0.11699580401182175,
-0.12227876484394073,
0.11282453685998917,
-0.008607556112110615,
0.05487091839313507,
0.0019345557084307075,
-0.05940266698598862,
0.1190459206700325,
-0.011249573901295662,
0.10915304720401764,
-0.0589299350976944,
-0.0329253226518631,
0.04332856833934784,
-0.04985667020082474,
0.022804871201515198,
-0.12402242422103882,
0.027077611535787582,
-0.04755053296685219,
-0.03654800355434418,
-0.043617211282253265,
0.034832220524549484,
-0.03675200790166855,
-0.06636807322502136,
-0.04232978820800781,
0.0248000118881464,
0.037313517183065414,
-0.013701836578547955,
0.13872301578521729,
0.014526233077049255,
0.14299680292606354,
0.13858695328235626,
0.0889226496219635,
-0.06302368640899658,
-0.029086172580718994,
0.0014898882945999503,
-0.03622199222445488,
0.06759503483772278,
-0.15006285905838013,
0.03898439183831215,
0.12571556866168976,
0.004996049217879772,
0.14975549280643463,
0.0819820687174797,
-0.02342674694955349,
0.008411030285060406,
0.06890203058719635,
-0.15844212472438812,
-0.10973291844129562,
-0.001020138501189649,
-0.030523601919412613,
-0.12493152171373367,
0.07138723880052567,
0.11392353475093842,
-0.07724443823099136,
0.008972173556685448,
-0.005928868893533945,
0.009677737951278687,
-0.04128974676132202,
0.17837920784950256,
0.06498932093381882,
0.04822166636586189,
-0.07270776480436325,
0.07779987901449203,
0.04430501163005829,
-0.07522151619195938,
0.017982834950089455,
0.04051190987229347,
-0.07983975112438202,
-0.03993218019604683,
0.05719459056854248,
0.19657103717327118,
-0.026500873267650604,
-0.05837640538811684,
-0.14217284321784973,
-0.12414282560348511,
0.05752115324139595,
0.19310036301612854,
0.1015230044722557,
0.009244979359209538,
-0.0407465361058712,
0.02295321226119995,
-0.12040913105010986,
0.11817211657762527,
0.025530783459544182,
0.08812250196933746,
-0.15503020584583282,
0.12197127938270569,
-0.003166693029925227,
0.005835792515426874,
-0.02511296421289444,
0.0533980056643486,
-0.12648648023605347,
-0.008106779307126999,
-0.12939110398292542,
-0.0057511297054588795,
-0.021352045238018036,
0.004955151583999395,
0.008199969306588173,
-0.04779666289687157,
-0.06358945369720459,
0.010732771828770638,
-0.09921436756849289,
-0.01684890314936638,
0.04158332943916321,
0.06871245056390762,
-0.12146934121847153,
-0.03792039677500725,
0.023298079147934914,
-0.06416113674640656,
0.06683740764856339,
0.010809250175952911,
0.02455774135887623,
0.05409993603825569,
-0.17380690574645996,
0.03410600870847702,
0.07515175640583038,
0.016016332432627678,
0.04754935950040817,
-0.08406995236873627,
-0.010849055834114552,
-0.003519645659253001,
0.04983530193567276,
0.020560255274176598,
0.07867547869682312,
-0.12608133256435394,
0.011518741026520729,
-0.033205609768629074,
-0.06680702418088913,
-0.04999776557087898,
0.021888049319386482,
0.08440057933330536,
-0.0014336990425363183,
0.19968794286251068,
-0.09745403379201889,
0.013787619769573212,
-0.20276124775409698,
0.010990669950842857,
0.003788095898926258,
-0.11093506962060928,
-0.1269329935312271,
-0.059081677347421646,
0.04800479859113693,
-0.06481306254863739,
0.15488024055957794,
0.020550571382045746,
0.010769807733595371,
0.03645927086472511,
-0.04107758030295372,
0.039359379559755325,
0.0327058881521225,
0.22751174867153168,
0.029378481209278107,
-0.04600120708346367,
0.01687941513955593,
0.03237639740109444,
0.11742069572210312,
0.08802219480276108,
0.1792479306459427,
0.16454298794269562,
-0.05248333141207695,
0.10774432122707367,
0.05408168584108353,
-0.05424882471561432,
-0.1418561339378357,
0.06450225412845612,
-0.047039780765771866,
0.10320526361465454,
-0.025574281811714172,
0.21490724384784698,
0.08548427373170853,
-0.1546553373336792,
0.007705057039856911,
-0.05943000689148903,
-0.08370780944824219,
-0.11372417211532593,
-0.050362713634967804,
-0.10065057128667831,
-0.1454731523990631,
0.0020896592177450657,
-0.11639350652694702,
0.0005258803721517324,
0.10065609216690063,
0.006369436159729958,
-0.0162113718688488,
0.1770777702331543,
0.0010083121014758945,
0.041153401136398315,
0.058675747364759445,
0.005907336715608835,
-0.030863314867019653,
-0.10153139382600784,
-0.08471550047397614,
-0.005867890082299709,
-0.020050017163157463,
0.01951914280653,
-0.05173894390463829,
-0.025494124740362167,
0.03894948959350586,
-0.005324534140527248,
-0.0950087457895279,
0.007959420792758465,
0.02015066333115101,
0.05255744606256485,
0.03345246613025665,
0.002984222024679184,
0.004778455942869186,
0.00007871716661611572,
0.19919419288635254,
-0.07997587323188782,
-0.058711644262075424,
-0.09603152424097061,
0.21290911734104156,
0.024051059037446976,
0.025115132331848145,
0.011564715765416622,
-0.08499964326620102,
0.01889016665518284,
0.2316461205482483,
0.18705515563488007,
-0.08112160861492157,
0.0009208113770000637,
0.0007524872780777514,
-0.01007313746958971,
-0.037841372191905975,
0.10160250216722488,
0.11669113487005234,
0.03306255489587784,
-0.07606016099452972,
-0.051873039454221725,
-0.031216835603117943,
0.0009899027645587921,
-0.03560937941074371,
0.0544857494533062,
0.04164080694317818,
0.012878463603556156,
-0.051990870386362076,
0.04565497115254402,
-0.03295009955763817,
-0.10391131043434143,
0.06071539223194122,
-0.19903860986232758,
-0.14849072694778442,
-0.004860726650804281,
0.12431804090738297,
-0.020456483587622643,
0.0531594455242157,
-0.03469376638531685,
-0.007731322664767504,
0.06413941830396652,
-0.024838242679834366,
-0.07736081629991531,
-0.06824623793363571,
0.05854307860136032,
-0.08722294121980667,
0.23744554817676544,
-0.03763803094625473,
0.05899600684642792,
0.13167360424995422,
0.04306875914335251,
-0.06833373755216599,
0.07767730206251144,
0.04729393124580383,
-0.07660970091819763,
0.025751251727342606,
0.06745623797178268,
-0.04449795186519623,
0.12284586578607559,
0.054649583995342255,
-0.13653752207756042,
0.01752297207713127,
-0.05634380504488945,
-0.08683975040912628,
-0.05936887860298157,
-0.036316003650426865,
-0.06538940966129303,
0.13406138122081757,
0.19432432949543,
-0.028924215584993362,
0.0025454042479395866,
-0.04353180527687073,
0.027164896950125694,
0.0671609416604042,
0.0414615124464035,
-0.03198600187897682,
-0.23235701024532318,
0.026354726403951645,
0.06970401108264923,
-0.00801156461238861,
-0.28303834795951843,
-0.09084466844797134,
0.0004814456042367965,
-0.04556996747851372,
-0.10062243044376373,
0.07513613998889923,
0.13156946003437042,
0.05771801248192787,
-0.06654103100299835,
-0.101446233689785,
-0.07540600746870041,
0.15229882299900055,
-0.13379502296447754,
-0.10312428325414658
] |
null | null | transformers | # dolly-v2-3b Model Card
## Summary
Databricks' `dolly-v2-3b`, an instruction-following large language model trained on the Databricks machine learning platform
that is licensed for commercial use. Based on `pythia-2.8b`, Dolly is trained on ~15k instruction/response fine tuning records
[`databricks-dolly-15k`](https://github.com/databrickslabs/dolly/tree/master/data) generated
by Databricks employees in capability domains from the InstructGPT paper, including brainstorming, classification, closed QA, generation,
information extraction, open QA and summarization. `dolly-v2-3b` is not a state-of-the-art model, but does exhibit surprisingly
high quality instruction following behavior not characteristic of the foundation model on which it is based.
Dolly v2 is also available in these larger models sizes:
* [dolly-v2-12b](https://huggingface.co/databricks/dolly-v2-12b), a 12 billion parameter based on `pythia-12b`
* [dolly-v2-7b](https://huggingface.co/databricks/dolly-v2-7b), a 6.9 billion parameter based on `pythia-6.9b`
Please refer to the [dolly GitHub repo](https://github.com/databrickslabs/dolly#getting-started-with-response-generation) for tips on
running inference for various GPU configurations.
**Owner**: Databricks, Inc.
## Model Overview
`dolly-v2-3b` is a 2.8 billion parameter causal language model created by [Databricks](https://databricks.com/) that is derived from
[EleutherAI's](https://www.eleuther.ai/) [Pythia-2.8b](https://huggingface.co/EleutherAI/pythia-2.8b) and fine-tuned
on a [~15K record instruction corpus](https://github.com/databrickslabs/dolly/tree/master/data) generated by Databricks employees and released under a permissive license (CC-BY-SA)
## Usage
To use the model with the `transformers` library on a machine with GPUs, first make sure you have the `transformers` and `accelerate` libraries installed.
In a Databricks notebook you could run:
```python
%pip install "accelerate>=0.16.0,<1" "transformers[torch]>=4.28.1,<5" "torch>=1.13.1,<2"
```
The instruction following pipeline can be loaded using the `pipeline` function as shown below. This loads a custom `InstructionTextGenerationPipeline`
found in the model repo [here](https://huggingface.co/databricks/dolly-v2-3b/blob/main/instruct_pipeline.py), which is why `trust_remote_code=True` is required.
Including `torch_dtype=torch.bfloat16` is generally recommended if this type is supported in order to reduce memory usage. It does not appear to impact output quality.
It is also fine to remove it if there is sufficient memory.
```python
import torch
from transformers import pipeline
generate_text = pipeline(model="databricks/dolly-v2-3b", torch_dtype=torch.bfloat16, trust_remote_code=True, device_map="auto")
```
You can then use the pipeline to answer instructions:
```python
res = generate_text("Explain to me the difference between nuclear fission and fusion.")
print(res[0]["generated_text"])
```
Alternatively, if you prefer to not use `trust_remote_code=True` you can download [instruct_pipeline.py](https://huggingface.co/databricks/dolly-v2-3b/blob/main/instruct_pipeline.py),
store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer:
```python
import torch
from instruct_pipeline import InstructionTextGenerationPipeline
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("databricks/dolly-v2-3b", padding_side="left")
model = AutoModelForCausalLM.from_pretrained("databricks/dolly-v2-3b", device_map="auto", torch_dtype=torch.bfloat16)
generate_text = InstructionTextGenerationPipeline(model=model, tokenizer=tokenizer)
```
### LangChain Usage
To use the pipeline with LangChain, you must set `return_full_text=True`, as LangChain expects the full text to be returned
and the default for the pipeline is to only return the new text.
```python
import torch
from transformers import pipeline
generate_text = pipeline(model="databricks/dolly-v2-3b", torch_dtype=torch.bfloat16,
trust_remote_code=True, device_map="auto", return_full_text=True)
```
You can create a prompt that either has only an instruction or has an instruction with context:
```python
from langchain import PromptTemplate, LLMChain
from langchain.llms import HuggingFacePipeline
# template for an instrution with no input
prompt = PromptTemplate(
input_variables=["instruction"],
template="{instruction}")
# template for an instruction with input
prompt_with_context = PromptTemplate(
input_variables=["instruction", "context"],
template="{instruction}\n\nInput:\n{context}")
hf_pipeline = HuggingFacePipeline(pipeline=generate_text)
llm_chain = LLMChain(llm=hf_pipeline, prompt=prompt)
llm_context_chain = LLMChain(llm=hf_pipeline, prompt=prompt_with_context)
```
Example predicting using a simple instruction:
```python
print(llm_chain.predict(instruction="Explain to me the difference between nuclear fission and fusion.").lstrip())
```
Example predicting using an instruction with context:
```python
context = """George Washington (February 22, 1732[b] - December 14, 1799) was an American military officer, statesman,
and Founding Father who served as the first president of the United States from 1789 to 1797."""
print(llm_context_chain.predict(instruction="When was George Washington president?", context=context).lstrip())
```
## Known Limitations
### Performance Limitations
**`dolly-v2-3b` is not a state-of-the-art generative language model** and, though quantitative benchmarking is ongoing, is not designed to perform
competitively with more modern model architectures or models subject to larger pretraining corpuses.
The Dolly model family is under active development, and so any list of shortcomings is unlikely to be exhaustive, but we include known limitations and misfires here as a means to document and share our preliminary findings with the community.
In particular, `dolly-v2-3b` struggles with: syntactically complex prompts, programming problems, mathematical operations, factual errors,
dates and times, open-ended question answering, hallucination, enumerating lists of specific length, stylistic mimicry, having a sense of humor, etc.
Moreover, we find that `dolly-v2-3b` does not have some capabilities, such as well-formatted letter writing, present in the original model.
### Dataset Limitations
Like all language models, `dolly-v2-3b` reflects the content and limitations of its training corpuses.
- **The Pile**: GPT-J's pre-training corpus contains content mostly collected from the public internet, and like most web-scale datasets,
it contains content many users would find objectionable. As such, the model is likely to reflect these shortcomings, potentially overtly
in the case it is explicitly asked to produce objectionable content, and sometimes subtly, as in the case of biased or harmful implicit
associations.
- **`databricks-dolly-15k`**: The training data on which `dolly-v2-3b` is instruction tuned represents natural language instructions generated
by Databricks employees during a period spanning March and April 2023 and includes passages from Wikipedia as references passages
for instruction categories like closed QA and summarization. To our knowledge it does not contain obscenity, intellectual property or
personally identifying information about non-public figures, but it may contain typos and factual errors.
The dataset may also reflect biases found in Wikipedia. Finally, the dataset likely reflects
the interests and semantic choices of Databricks employees, a demographic which is not representative of the global population at large.
Databricks is committed to ongoing research and development efforts to develop helpful, honest and harmless AI technologies that
maximize the potential of all individuals and organizations.
### Benchmark Metrics
Below you'll find various models benchmark performance on the [EleutherAI LLM Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness);
model results are sorted by geometric mean to produce an intelligible ordering. As outlined above, these results demonstrate that `dolly-v2-3b` is not state of the art.
It underperforms `dolly-v1-6b` in the evaluation benchmarks, which is not surprising considering it has half the number of parameters.
| model | openbookqa | arc_easy | winogrande | hellaswag | arc_challenge | piqa | boolq | gmean |
| --------------------------------- | ------------ | ---------- | ------------ | ----------- | --------------- | -------- | -------- | ---------|
| EleutherAI/pythia-2.8b | 0.348 | 0.585859 | 0.589582 | 0.591217 | 0.323379 | 0.73395 | 0.638226 | 0.523431 |
| EleutherAI/pythia-6.9b | 0.368 | 0.604798 | 0.608524 | 0.631548 | 0.343857 | 0.761153 | 0.6263 | 0.543567 |
| databricks/dolly-v2-3b | 0.384 | 0.611532 | 0.589582 | 0.650767 | 0.370307 | 0.742655 | 0.575535 | 0.544886 |
| EleutherAI/pythia-12b | 0.364 | 0.627104 | 0.636148 | 0.668094 | 0.346416 | 0.760065 | 0.673394 | 0.559676 |
| EleutherAI/gpt-j-6B | 0.382 | 0.621633 | 0.651144 | 0.662617 | 0.363481 | 0.761153 | 0.655963 | 0.565936 |
| databricks/dolly-v2-12b | 0.408 | 0.63931 | 0.616417 | 0.707927 | 0.388225 | 0.757889 | 0.568196 | 0.56781 |
| databricks/dolly-v2-7b | 0.392 | 0.633838 | 0.607735 | 0.686517 | 0.406997 | 0.750816 | 0.644037 | 0.573487 |
| databricks/dolly-v1-6b | 0.41 | 0.62963 | 0.643252 | 0.676758 | 0.384812 | 0.773667 | 0.687768 | 0.583431 |
| EleutherAI/gpt-neox-20b | 0.402 | 0.683923 | 0.656669 | 0.7142 | 0.408703 | 0.784004 | 0.695413 | 0.602236 |
# Citation
```
@online{DatabricksBlog2023DollyV2,
author = {Mike Conover and Matt Hayes and Ankit Mathur and Jianwei Xie and Jun Wan and Sam Shah and Ali Ghodsi and Patrick Wendell and Matei Zaharia and Reynold Xin},
title = {Free Dolly: Introducing the World's First Truly Open Instruction-Tuned LLM},
year = {2023},
url = {https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm},
urldate = {2023-06-30}
}
```
# Happy Hacking! | {"language": ["en"], "license": "mit", "library_name": "transformers", "datasets": ["databricks/databricks-dolly-15k"], "inference": false} | text-generation | BashitAli/GPT_model | [
"transformers",
"pytorch",
"gpt_neox",
"text-generation",
"en",
"dataset:databricks/databricks-dolly-15k",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T06:59:34+00:00 | [] | [
"en"
] | TAGS
#transformers #pytorch #gpt_neox #text-generation #en #dataset-databricks/databricks-dolly-15k #license-mit #autotrain_compatible #text-generation-inference #region-us
| dolly-v2-3b Model Card
======================
Summary
-------
Databricks' 'dolly-v2-3b', an instruction-following large language model trained on the Databricks machine learning platform
that is licensed for commercial use. Based on 'pythia-2.8b', Dolly is trained on ~15k instruction/response fine tuning records
'databricks-dolly-15k' generated
by Databricks employees in capability domains from the InstructGPT paper, including brainstorming, classification, closed QA, generation,
information extraction, open QA and summarization. 'dolly-v2-3b' is not a state-of-the-art model, but does exhibit surprisingly
high quality instruction following behavior not characteristic of the foundation model on which it is based.
Dolly v2 is also available in these larger models sizes:
* dolly-v2-12b, a 12 billion parameter based on 'pythia-12b'
* dolly-v2-7b, a 6.9 billion parameter based on 'pythia-6.9b'
Please refer to the dolly GitHub repo for tips on
running inference for various GPU configurations.
Owner: Databricks, Inc.
Model Overview
--------------
'dolly-v2-3b' is a 2.8 billion parameter causal language model created by Databricks that is derived from
EleutherAI's Pythia-2.8b and fine-tuned
on a ~15K record instruction corpus generated by Databricks employees and released under a permissive license (CC-BY-SA)
Usage
-----
To use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' and 'accelerate' libraries installed.
In a Databricks notebook you could run:
The instruction following pipeline can be loaded using the 'pipeline' function as shown below. This loads a custom 'InstructionTextGenerationPipeline'
found in the model repo here, which is why 'trust\_remote\_code=True' is required.
Including 'torch\_dtype=torch.bfloat16' is generally recommended if this type is supported in order to reduce memory usage. It does not appear to impact output quality.
It is also fine to remove it if there is sufficient memory.
You can then use the pipeline to answer instructions:
Alternatively, if you prefer to not use 'trust\_remote\_code=True' you can download instruct\_pipeline.py,
store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer:
### LangChain Usage
To use the pipeline with LangChain, you must set 'return\_full\_text=True', as LangChain expects the full text to be returned
and the default for the pipeline is to only return the new text.
You can create a prompt that either has only an instruction or has an instruction with context:
Example predicting using a simple instruction:
Example predicting using an instruction with context:
Known Limitations
-----------------
### Performance Limitations
'dolly-v2-3b' is not a state-of-the-art generative language model and, though quantitative benchmarking is ongoing, is not designed to perform
competitively with more modern model architectures or models subject to larger pretraining corpuses.
The Dolly model family is under active development, and so any list of shortcomings is unlikely to be exhaustive, but we include known limitations and misfires here as a means to document and share our preliminary findings with the community.
In particular, 'dolly-v2-3b' struggles with: syntactically complex prompts, programming problems, mathematical operations, factual errors,
dates and times, open-ended question answering, hallucination, enumerating lists of specific length, stylistic mimicry, having a sense of humor, etc.
Moreover, we find that 'dolly-v2-3b' does not have some capabilities, such as well-formatted letter writing, present in the original model.
### Dataset Limitations
Like all language models, 'dolly-v2-3b' reflects the content and limitations of its training corpuses.
* The Pile: GPT-J's pre-training corpus contains content mostly collected from the public internet, and like most web-scale datasets,
it contains content many users would find objectionable. As such, the model is likely to reflect these shortcomings, potentially overtly
in the case it is explicitly asked to produce objectionable content, and sometimes subtly, as in the case of biased or harmful implicit
associations.
* 'databricks-dolly-15k': The training data on which 'dolly-v2-3b' is instruction tuned represents natural language instructions generated
by Databricks employees during a period spanning March and April 2023 and includes passages from Wikipedia as references passages
for instruction categories like closed QA and summarization. To our knowledge it does not contain obscenity, intellectual property or
personally identifying information about non-public figures, but it may contain typos and factual errors.
The dataset may also reflect biases found in Wikipedia. Finally, the dataset likely reflects
the interests and semantic choices of Databricks employees, a demographic which is not representative of the global population at large.
Databricks is committed to ongoing research and development efforts to develop helpful, honest and harmless AI technologies that
maximize the potential of all individuals and organizations.
### Benchmark Metrics
Below you'll find various models benchmark performance on the EleutherAI LLM Evaluation Harness;
model results are sorted by geometric mean to produce an intelligible ordering. As outlined above, these results demonstrate that 'dolly-v2-3b' is not state of the art.
It underperforms 'dolly-v1-6b' in the evaluation benchmarks, which is not surprising considering it has half the number of parameters.
Happy Hacking!
==============
| [
"### LangChain Usage\n\n\nTo use the pipeline with LangChain, you must set 'return\\_full\\_text=True', as LangChain expects the full text to be returned\nand the default for the pipeline is to only return the new text.\n\n\nYou can create a prompt that either has only an instruction or has an instruction with context:\n\n\nExample predicting using a simple instruction:\n\n\nExample predicting using an instruction with context:\n\n\nKnown Limitations\n-----------------",
"### Performance Limitations\n\n\n'dolly-v2-3b' is not a state-of-the-art generative language model and, though quantitative benchmarking is ongoing, is not designed to perform\ncompetitively with more modern model architectures or models subject to larger pretraining corpuses.\n\n\nThe Dolly model family is under active development, and so any list of shortcomings is unlikely to be exhaustive, but we include known limitations and misfires here as a means to document and share our preliminary findings with the community. \n\nIn particular, 'dolly-v2-3b' struggles with: syntactically complex prompts, programming problems, mathematical operations, factual errors,\ndates and times, open-ended question answering, hallucination, enumerating lists of specific length, stylistic mimicry, having a sense of humor, etc.\nMoreover, we find that 'dolly-v2-3b' does not have some capabilities, such as well-formatted letter writing, present in the original model.",
"### Dataset Limitations\n\n\nLike all language models, 'dolly-v2-3b' reflects the content and limitations of its training corpuses.\n\n\n* The Pile: GPT-J's pre-training corpus contains content mostly collected from the public internet, and like most web-scale datasets,\nit contains content many users would find objectionable. As such, the model is likely to reflect these shortcomings, potentially overtly\nin the case it is explicitly asked to produce objectionable content, and sometimes subtly, as in the case of biased or harmful implicit\nassociations.\n* 'databricks-dolly-15k': The training data on which 'dolly-v2-3b' is instruction tuned represents natural language instructions generated\nby Databricks employees during a period spanning March and April 2023 and includes passages from Wikipedia as references passages\nfor instruction categories like closed QA and summarization. To our knowledge it does not contain obscenity, intellectual property or\npersonally identifying information about non-public figures, but it may contain typos and factual errors.\nThe dataset may also reflect biases found in Wikipedia. Finally, the dataset likely reflects\nthe interests and semantic choices of Databricks employees, a demographic which is not representative of the global population at large.\n\n\nDatabricks is committed to ongoing research and development efforts to develop helpful, honest and harmless AI technologies that\nmaximize the potential of all individuals and organizations.",
"### Benchmark Metrics\n\n\nBelow you'll find various models benchmark performance on the EleutherAI LLM Evaluation Harness;\nmodel results are sorted by geometric mean to produce an intelligible ordering. As outlined above, these results demonstrate that 'dolly-v2-3b' is not state of the art.\nIt underperforms 'dolly-v1-6b' in the evaluation benchmarks, which is not surprising considering it has half the number of parameters.\n\n\n\nHappy Hacking!\n=============="
] | [
"TAGS\n#transformers #pytorch #gpt_neox #text-generation #en #dataset-databricks/databricks-dolly-15k #license-mit #autotrain_compatible #text-generation-inference #region-us \n",
"### LangChain Usage\n\n\nTo use the pipeline with LangChain, you must set 'return\\_full\\_text=True', as LangChain expects the full text to be returned\nand the default for the pipeline is to only return the new text.\n\n\nYou can create a prompt that either has only an instruction or has an instruction with context:\n\n\nExample predicting using a simple instruction:\n\n\nExample predicting using an instruction with context:\n\n\nKnown Limitations\n-----------------",
"### Performance Limitations\n\n\n'dolly-v2-3b' is not a state-of-the-art generative language model and, though quantitative benchmarking is ongoing, is not designed to perform\ncompetitively with more modern model architectures or models subject to larger pretraining corpuses.\n\n\nThe Dolly model family is under active development, and so any list of shortcomings is unlikely to be exhaustive, but we include known limitations and misfires here as a means to document and share our preliminary findings with the community. \n\nIn particular, 'dolly-v2-3b' struggles with: syntactically complex prompts, programming problems, mathematical operations, factual errors,\ndates and times, open-ended question answering, hallucination, enumerating lists of specific length, stylistic mimicry, having a sense of humor, etc.\nMoreover, we find that 'dolly-v2-3b' does not have some capabilities, such as well-formatted letter writing, present in the original model.",
"### Dataset Limitations\n\n\nLike all language models, 'dolly-v2-3b' reflects the content and limitations of its training corpuses.\n\n\n* The Pile: GPT-J's pre-training corpus contains content mostly collected from the public internet, and like most web-scale datasets,\nit contains content many users would find objectionable. As such, the model is likely to reflect these shortcomings, potentially overtly\nin the case it is explicitly asked to produce objectionable content, and sometimes subtly, as in the case of biased or harmful implicit\nassociations.\n* 'databricks-dolly-15k': The training data on which 'dolly-v2-3b' is instruction tuned represents natural language instructions generated\nby Databricks employees during a period spanning March and April 2023 and includes passages from Wikipedia as references passages\nfor instruction categories like closed QA and summarization. To our knowledge it does not contain obscenity, intellectual property or\npersonally identifying information about non-public figures, but it may contain typos and factual errors.\nThe dataset may also reflect biases found in Wikipedia. Finally, the dataset likely reflects\nthe interests and semantic choices of Databricks employees, a demographic which is not representative of the global population at large.\n\n\nDatabricks is committed to ongoing research and development efforts to develop helpful, honest and harmless AI technologies that\nmaximize the potential of all individuals and organizations.",
"### Benchmark Metrics\n\n\nBelow you'll find various models benchmark performance on the EleutherAI LLM Evaluation Harness;\nmodel results are sorted by geometric mean to produce an intelligible ordering. As outlined above, these results demonstrate that 'dolly-v2-3b' is not state of the art.\nIt underperforms 'dolly-v1-6b' in the evaluation benchmarks, which is not surprising considering it has half the number of parameters.\n\n\n\nHappy Hacking!\n=============="
] | [
64,
109,
227,
332,
119
] | [
"passage: TAGS\n#transformers #pytorch #gpt_neox #text-generation #en #dataset-databricks/databricks-dolly-15k #license-mit #autotrain_compatible #text-generation-inference #region-us \n### LangChain Usage\n\n\nTo use the pipeline with LangChain, you must set 'return\\_full\\_text=True', as LangChain expects the full text to be returned\nand the default for the pipeline is to only return the new text.\n\n\nYou can create a prompt that either has only an instruction or has an instruction with context:\n\n\nExample predicting using a simple instruction:\n\n\nExample predicting using an instruction with context:\n\n\nKnown Limitations\n-----------------### Performance Limitations\n\n\n'dolly-v2-3b' is not a state-of-the-art generative language model and, though quantitative benchmarking is ongoing, is not designed to perform\ncompetitively with more modern model architectures or models subject to larger pretraining corpuses.\n\n\nThe Dolly model family is under active development, and so any list of shortcomings is unlikely to be exhaustive, but we include known limitations and misfires here as a means to document and share our preliminary findings with the community. \n\nIn particular, 'dolly-v2-3b' struggles with: syntactically complex prompts, programming problems, mathematical operations, factual errors,\ndates and times, open-ended question answering, hallucination, enumerating lists of specific length, stylistic mimicry, having a sense of humor, etc.\nMoreover, we find that 'dolly-v2-3b' does not have some capabilities, such as well-formatted letter writing, present in the original model."
] | [
-0.04902537539601326,
0.02031327225267887,
-0.004264208488166332,
0.06234452500939369,
0.0388268418610096,
-0.00886545144021511,
0.07225928455591202,
0.05958810821175575,
0.03370939940214157,
0.07986266911029816,
0.0326484851539135,
-0.030458632856607437,
0.04108487814664841,
-0.046566907316446304,
0.03471419960260391,
-0.25826969742774963,
0.02579585835337639,
-0.04863527789711952,
0.07920879870653152,
0.0343746580183506,
0.12029419839382172,
-0.03386799246072769,
0.09186465293169022,
0.0372178889811039,
-0.01577041670680046,
0.012602037750184536,
-0.005025828722864389,
0.028984755277633667,
0.13987240195274353,
0.07722001522779465,
0.09280440211296082,
-0.04700516536831856,
-0.022056542336940765,
-0.09139895439147949,
0.024620210751891136,
0.085695780813694,
0.010219770483672619,
0.026919476687908173,
0.011133302934467793,
0.02497098408639431,
0.23130497336387634,
-0.031430892646312714,
0.03072519600391388,
0.03618887811899185,
-0.111451655626297,
-0.08958280086517334,
-0.006159725598990917,
0.029583167284727097,
0.06535643339157104,
0.010908572003245354,
-0.012683580629527569,
0.09718015044927597,
-0.07673880457878113,
0.07011215388774872,
0.14130499958992004,
-0.21008001267910004,
-0.06727108359336853,
0.09874215722084045,
0.16092225909233093,
0.09023561328649521,
-0.03387364745140076,
-0.004445184953510761,
0.004134484566748142,
0.060212232172489166,
-0.04265459254384041,
-0.0527898333966732,
-0.0504070483148098,
0.00003896790440194309,
-0.15113432705402374,
-0.022087134420871735,
0.21950426697731018,
-0.021029463037848473,
-0.031523313373327255,
-0.14150889217853546,
-0.07554148137569427,
0.07292135804891586,
-0.027791285887360573,
-0.05577811226248741,
-0.006962483283132315,
0.047885823994874954,
0.14535145461559296,
-0.10106375068426132,
-0.11498334258794785,
-0.034942422062158585,
-0.03585614636540413,
0.06172759830951691,
0.06540390104055405,
0.052875641733407974,
-0.06754033267498016,
0.05401373282074928,
-0.12093936651945114,
-0.02546035125851631,
-0.018065448850393295,
-0.11013985425233841,
-0.08044388145208359,
0.010268101468682289,
-0.10658646374940872,
-0.0466286726295948,
0.04828159138560295,
0.05969448760151863,
0.0085047772154212,
0.007012657355517149,
0.038949038833379745,
0.06196430325508118,
0.07249479740858078,
0.07117629796266556,
-0.042958781123161316,
0.03716162592172623,
0.013302966021001339,
0.08046721667051315,
0.038289327174425125,
-0.00042077802936546504,
-0.06523120403289795,
-0.05232209339737892,
0.006860814522951841,
0.05506323650479317,
0.02135409228503704,
0.060077402740716934,
-0.06092546880245209,
0.003600317519158125,
0.06934208422899246,
-0.07986200600862503,
-0.017703447490930557,
-0.02550642378628254,
-0.030940502882003784,
-0.0902244970202446,
0.02714093215763569,
-0.01276144478470087,
-0.0973726212978363,
-0.015425545163452625,
-0.08044516295194626,
-0.04873232543468475,
-0.10124683380126953,
-0.09926170110702515,
0.025637131184339523,
-0.03671042621135712,
-0.05606542155146599,
-0.1012316420674324,
-0.1610749363899231,
0.0005789262359030545,
0.020326649770140648,
0.0032935587223619223,
-0.048263952136039734,
0.026743410155177116,
-0.02577461302280426,
-0.1013370081782341,
0.010692101903259754,
0.007608870044350624,
-0.02070206217467785,
-0.006333512719720602,
-0.04925638809800148,
0.03965237736701965,
-0.0632438063621521,
-0.008141841739416122,
-0.07930786162614822,
0.06771258264780045,
-0.14957450330257416,
0.09221398085355759,
-0.04628190025687218,
-0.07342921942472458,
-0.03728717193007469,
0.005546254571527243,
-0.05691447854042053,
0.08969827741384506,
-0.0052396091632544994,
0.07270347326993942,
-0.15388458967208862,
0.027026992291212082,
0.18179666996002197,
-0.1003841757774353,
-0.061579786241054535,
0.1811293363571167,
-0.04328000918030739,
0.07176347821950912,
0.12384945899248123,
0.07544159889221191,
0.12975941598415375,
-0.12916439771652222,
-0.022852573543787003,
0.06056753918528557,
-0.08663841336965561,
0.11865319311618805,
0.03961430490016937,
-0.02209439128637314,
-0.060246508568525314,
0.03914427012205124,
-0.06545428186655045,
-0.02600765973329544,
-0.01060317363590002,
-0.03755740448832512,
-0.022405283525586128,
-0.020959757268428802,
0.0196970384567976,
0.004774000030010939,
-0.03491272032260895,
-0.012994633987545967,
-0.0927390456199646,
-0.0329362116754055,
0.09431271255016327,
-0.020476868376135826,
0.0064382050186395645,
-0.09618915617465973,
0.01982191763818264,
0.003507704706862569,
0.01406601257622242,
-0.1934799700975418,
-0.10746074467897415,
0.08413910865783691,
-0.11208553612232208,
0.0692913755774498,
0.15408077836036682,
0.020404990762472153,
0.07256238907575607,
0.018106810748577118,
0.04872877150774002,
-0.017640715464949608,
-0.04073439911007881,
-0.09940386563539505,
-0.12118835002183914,
-0.05189090967178345,
-0.03257782757282257,
0.14010655879974365,
-0.1353873908519745,
0.03235182166099548,
-0.03763704374432564,
0.00770595483481884,
0.011234413832426071,
-0.031772468239068985,
-0.04907890036702156,
0.016953807324171066,
0.005092248320579529,
0.009913071990013123,
0.06483791768550873,
0.03545273095369339,
-0.05186304450035095,
0.1048557385802269,
-0.16450920701026917,
-0.14278602600097656,
0.03260133042931557,
0.005844105500727892,
-0.05168665200471878,
-0.04422454163432121,
-0.04762166365981102,
-0.04007667303085327,
-0.039840664714574814,
-0.13868363201618195,
0.20327718555927277,
0.02784097194671631,
0.07132185995578766,
-0.13197878003120422,
-0.034339696168899536,
-0.022410040721297264,
-0.06723658740520477,
0.023090653121471405,
0.08888401836156845,
0.09437616914510727,
-0.08877276629209518,
0.0009214659221470356,
-0.023823730647563934,
-0.097429059445858,
0.13437457382678986,
0.013539282605051994,
-0.05179569497704506,
-0.03140239417552948,
0.051299698650836945,
-0.0025223898701369762,
0.06909311562776566,
-0.12828023731708527,
-0.011487023904919624,
0.02713816799223423,
-0.004988979548215866,
0.06821194291114807,
-0.14238715171813965,
0.03861397132277489,
0.0036292269360274076,
-0.00539738405495882,
0.09320685267448425,
0.015825143083930016,
0.03558286651968956,
0.1328103244304657,
-0.014943747781217098,
-0.02030404657125473,
-0.03609032928943634,
-0.0919196680188179,
-0.11398911476135254,
0.15860515832901,
-0.0795876681804657,
-0.192367285490036,
-0.12679435312747955,
0.03959580883383751,
-0.10612879693508148,
0.0014637003187090158,
0.01893259957432747,
0.004241427406668663,
-0.04933324456214905,
-0.11514195054769516,
0.11268399655818939,
-0.06578879058361053,
-0.0948575884103775,
-0.08682626485824585,
0.008181269280612469,
-0.039963822811841965,
-0.15648412704467773,
-0.01606939546763897,
-0.0447830893099308,
-0.10361744463443756,
-0.00011815741163445637,
-0.02097455970942974,
0.04494715854525566,
0.07682474702596664,
0.058506079018116,
-0.07939323782920837,
-0.0506506972014904,
0.18180686235427856,
-0.0066175153478980064,
0.08245448023080826,
0.20494765043258667,
-0.049533627927303314,
0.09964954853057861,
0.07946991175413132,
0.015019885264337063,
-0.05473299324512482,
0.02991345338523388,
0.047984931617975235,
-0.04443207010626793,
-0.15669459104537964,
-0.1255558729171753,
-0.0932178944349289,
0.021220171824097633,
0.053093671798706055,
0.039886198937892914,
0.01640239916741848,
0.032113369554281235,
-0.0781819298863411,
0.06303510069847107,
0.03746483102440834,
0.06192350387573242,
0.1595456451177597,
-0.050056908279657364,
0.05552788823843002,
-0.051803167909383774,
-0.055055614560842514,
0.07685121148824692,
0.05342429503798485,
0.2554282546043396,
-0.04465709999203682,
0.16429969668388367,
0.11569816619157791,
0.06315045803785324,
0.03426390886306763,
-0.019226713106036186,
-0.010875890031456947,
0.047738343477249146,
-0.05851313844323158,
-0.07196418941020966,
-0.03826916217803955,
0.10715270787477493,
0.02884025126695633,
-0.038370538502931595,
-0.036053258925676346,
0.05909816175699234,
0.03335488960146904,
0.2507583498954773,
-0.03369935229420662,
-0.1783464103937149,
-0.07068731635808945,
0.03859550878405571,
-0.0460551492869854,
-0.08200399577617645,
0.026936020702123642,
0.12213753908872604,
-0.10551615804433823,
-0.03868044167757034,
-0.01436181552708149,
0.061640217900276184,
-0.033274952322244644,
0.04223326966166496,
-0.06044755131006241,
0.14547766745090485,
0.04483705386519432,
0.10338466614484787,
-0.1391983926296234,
0.11744223535060883,
0.0033514464739710093,
0.043254777789115906,
-0.10226409137248993,
0.03783666342496872,
0.02530747279524803,
-0.005277504678815603,
0.08770734816789627,
0.015950065106153488,
0.026285767555236816,
0.005812990013509989,
0.0004072446608915925,
0.04624427855014801,
0.07066353410482407,
-0.014749766327440739,
0.04628346115350723,
-0.05214773491024971,
0.03862619027495384,
0.0005610272055491805,
-0.007898607291281223,
-0.098551906645298,
-0.16176943480968475,
0.03255639970302582,
-0.010183342732489109,
-0.09757675975561142,
-0.08240599930286407,
-0.0029353953432291746,
0.021172761917114258,
0.10541462153196335,
-0.0688743144273758,
-0.08881325274705887,
-0.041471391916275024,
-0.0410759411752224,
0.08863171935081482,
-0.05864691734313965,
-0.034833889454603195,
-0.06111786141991615,
0.11689341813325882,
-0.06370146572589874,
-0.05615658313035965,
0.037441570311784744,
-0.07586649805307388,
-0.11837967485189438,
-0.07060269266366959,
0.11384362727403641,
0.10970919579267502,
0.019950054585933685,
-0.015063579194247723,
-0.014062791131436825,
-0.03545713424682617,
-0.1089002937078476,
-0.04193725809454918,
0.17579086124897003,
0.05101674050092697,
0.10716760903596878,
-0.0950331762433052,
0.0032788123935461044,
-0.06319347769021988,
0.024521905928850174,
0.10551318526268005,
0.24480198323726654,
-0.05637776106595993,
0.06413507461547852,
0.1621505171060562,
-0.11488527804613113,
-0.2550506293773651,
-0.05422045662999153,
0.06322228163480759,
-0.008579644374549389,
0.04895106703042984,
-0.11939487606287003,
0.10899233818054199,
0.10845271497964859,
-0.021005265414714813,
0.009629867970943451,
-0.35896146297454834,
-0.11411152780056,
0.05161823704838753,
0.027230970561504364,
0.09547805786132812,
-0.12335649877786636,
0.011099845170974731,
0.036537472158670425,
-0.0319453626871109,
0.09653083235025406,
0.0028832892421633005,
0.06709616631269455,
0.0043914467096328735,
-0.01074502244591713,
0.041299350559711456,
-0.009630250744521618,
0.16579602658748627,
-0.03798651695251465,
0.06275621056556702,
-0.04476050287485123,
-0.020530663430690765,
0.09629672020673752,
-0.05961218103766441,
0.12115728110074997,
0.004900562111288309,
0.011250164359807968,
-0.14092271029949188,
-0.020841460675001144,
-0.0937313362956047,
0.08970613777637482,
-0.03697320818901062,
-0.06444988399744034,
-0.10454276949167252,
0.05617213994264603,
0.024516168981790543,
0.035209447145462036,
-0.05094475299119949,
-0.06799207627773285,
0.0784054771065712,
0.07808683812618256,
0.18630099296569824,
-0.001246095634996891,
-0.11060531437397003,
0.018079940229654312,
0.03302527219057083,
0.10821008682250977,
-0.027862217277288437,
0.03134531155228615,
0.07887106388807297,
-0.02368609420955181,
0.09987949579954147,
0.027918608859181404,
-0.11886413395404816,
0.011094911955296993,
0.023792343214154243,
-0.13112831115722656,
-0.17295104265213013,
-0.03517308086156845,
0.06632081419229507,
-0.08747676014900208,
-0.03589089959859848,
0.11025728285312653,
-0.06574924290180206,
-0.012318634428083897,
-0.015187949873507023,
0.05659420043230057,
-0.0012949192896485329,
-0.0017086518928408623,
-0.01970626413822174,
0.03745925799012184,
-0.033419352024793625,
0.08936844021081924,
0.08893812447786331,
-0.14163732528686523,
0.04745320975780487,
0.10135349631309509,
-0.03066529519855976,
-0.060029786080121994,
-0.12596717476844788,
0.11831602454185486,
0.02850181981921196,
-0.07235284149646759,
-0.0003684737312141806,
-0.09225541353225708,
0.031593360006809235,
0.11071408540010452,
0.016176044940948486,
0.029800178483128548,
-0.036293212324380875,
0.053922202438116074,
-0.0729399025440216,
0.0634303092956543,
-0.008081112988293171,
0.03635985776782036,
-0.034185025840997696,
0.06458853930234909,
0.009313352406024933,
0.037680551409721375,
-0.014390675351023674,
-0.05978010222315788,
-0.11412706226110458,
-0.009899521246552467,
-0.20494095981121063,
0.027179019525647163,
-0.07846029847860336,
0.013847949914634228,
0.010693551041185856,
0.007281463127583265,
-0.028070418164134026,
0.007089214399456978,
-0.04704786837100983,
0.007483485154807568,
-0.02445666864514351,
0.12225484102964401,
-0.11655839532613754,
0.003011665539816022,
0.07696148008108139,
-0.034682441502809525,
0.07209579646587372,
-0.03134164586663246,
-0.049007855355739594,
0.007848898880183697,
-0.00712839188054204,
0.05191219970583916,
-0.013254393823444843,
0.08882183581590652,
-0.0396045558154583,
-0.04347318783402443,
0.007959929294884205,
0.01291738636791706,
-0.004617806524038315,
-0.02566884644329548,
0.06876633316278458,
-0.1099814921617508,
0.07334136217832565,
0.08657843619585037,
-0.03743528947234154,
-0.09440255165100098,
-0.01837863400578499,
0.023482680320739746,
0.08437273651361465,
0.07605183869600296,
-0.024965820834040642,
-0.025516796857118607,
-0.16622792184352875,
-0.008868809789419174,
0.032979197800159454,
0.006379112135618925,
0.06270449608564377,
-0.09162762761116028,
0.04512948915362358,
-0.01295157615095377,
0.24854640662670135,
0.05209147930145264,
-0.019260907545685768,
0.03863174840807915,
0.05316407233476639,
-0.004206322133541107,
0.005754787474870682,
-0.049693115055561066,
-0.014247097074985504,
0.0017626432236284018,
-0.047687456011772156,
-0.025553615763783455,
-0.01528140902519226,
-0.033151641488075256,
0.1474967747926712,
0.07808912545442581,
0.13494285941123962,
0.02436554618179798,
0.09818226099014282,
0.0011870269663631916,
-0.004050104413181543,
0.02319924719631672,
0.0378665030002594,
0.04722971469163895,
-0.05971423536539078,
0.058702822774648666,
0.13036508858203888,
-0.0651450827717781,
0.13130876421928406,
-0.06422829627990723,
-0.04434996470808983,
-0.07968948781490326,
-0.16524727642536163,
-0.01157920341938734,
-0.003495562821626663,
-0.04625857621431351,
-0.13995321094989777,
-0.0018561143660917878,
0.08946945518255234,
-0.012623077258467674,
-0.03847000002861023,
0.07747144997119904,
-0.13900943100452423,
-0.07991770654916763,
0.020544331520795822,
-0.04850374907255173,
0.12146034091711044,
0.07352888584136963,
-0.038022466003894806,
0.029973141849040985,
0.04468824714422226,
0.026330027729272842,
0.06566925346851349,
0.008675585500895977,
-0.011946856044232845,
-0.08814427256584167,
-0.04060664772987366,
-0.00849228911101818,
-0.020867416635155678,
-0.04570290446281433,
0.11079630255699158,
0.06821265816688538,
-0.07885288447141647,
0.031160391867160797,
0.1908256560564041,
-0.020832959562540054,
-0.08003196865320206,
-0.14035092294216156,
0.21867014467716217,
-0.06234092637896538,
-0.022428303956985474,
0.035975679755210876,
-0.04823844134807587,
-0.0044272784143686295,
0.20894061028957367,
0.1974671185016632,
-0.08472593873739243,
-0.005085857585072517,
-0.004369736183434725,
0.00014269011444412172,
-0.02934933453798294,
0.2071627527475357,
0.041191112250089645,
0.35399308800697327,
-0.07167680561542511,
0.10757859796285629,
-0.021016620099544525,
0.05639783665537834,
-0.15232554078102112,
0.06123264133930206,
0.007601306773722172,
-0.0019439971074461937,
-0.018249519169330597,
0.06389178335666656,
-0.06075533106923103,
-0.19417844712734222,
-0.08915629237890244,
-0.027577703818678856,
-0.09707073867321014,
-0.0011173207312822342,
0.07341181486845016,
-0.035830605775117874,
0.10500424355268478,
-0.04911249130964279,
-0.0024479045532643795,
0.0808120146393776,
-0.02840293012559414,
-0.05537731945514679,
-0.02695411629974842,
0.12592625617980957,
-0.008653838187456131,
0.08793730288743973,
0.003661088878288865,
0.12125472724437714,
0.09162738919258118,
-0.029238468036055565,
-0.042624376714229584,
0.060211729258298874,
0.04146215692162514,
-0.053576163947582245,
-0.00782139040529728,
0.07235759496688843,
0.012796314433217049,
0.08479966968297958,
0.09272710233926773,
-0.03355766832828522,
0.05056134983897209,
0.055519264191389084,
0.021436216309666634,
-0.09771624207496643,
0.09939554333686829,
-0.11610889434814453,
0.13011649250984192,
0.1494092494249344,
0.0004099592624697834,
-0.030805554240942,
-0.06985671818256378,
-0.03877127543091774,
0.0029557484667748213,
0.08082442730665207,
-0.05976384878158569,
-0.10488968342542648,
0.004731060937047005,
0.029550805687904358,
0.07144384831190109,
-0.1551472544670105,
-0.0724402517080307,
0.03173041716217995,
0.0038939460646361113,
-0.016425734385848045,
0.056722111999988556,
0.08027733117341995,
0.0241034384816885,
-0.026085760444402695,
-0.14152945578098297,
-0.004779171198606491,
0.08008304983377457,
-0.09680518507957458,
-0.08361361175775528
] |
null | null | null |
<!-- header start -->
<!-- 200823 -->
<div style="width: auto; margin-left: auto; margin-right: auto">
<img src="https://github.com/second-state/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
</div>
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
<!-- header end -->
# Qwen1.5-7B-Chat-GGUF
## Original Model
[Qwen/Qwen1.5-7B-Chat](https://huggingface.co/Qwen/Qwen1.5-7B-Chat)
## Run with LlamaEdge
- LlamaEdge version: [v0.2.15](https://github.com/second-state/LlamaEdge/releases/tag/0.2.15) and above
- Prompt template
- Prompt type: `chatml`
- Prompt string
```text
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
```
- Run as LlamaEdge service
```bash
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Qwen1.5-7B-Chat-Q5_K_M.gguf llama-api-server.wasm -p chatml
```
- Run as LlamaEdge command app
```bash
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Qwen1.5-7B-Chat-Q5_K_M.gguf llama-chat.wasm -p chatml
```
## Quantized GGUF Models
| Name | Quant method | Bits | Size | Use case |
| ---- | ---- | ---- | ---- | ----- |
| [Qwen1.5-7B-Chat-Q2_K.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q2_K.gguf) | Q2_K | 2 | 3.10 GB| smallest, significant quality loss - not recommended for most purposes |
| [Qwen1.5-7B-Chat-Q3_K_L.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q3_K_L.gguf) | Q3_K_L | 3 | 4.22 GB| small, substantial quality loss |
| [Qwen1.5-7B-Chat-Q3_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q3_K_M.gguf) | Q3_K_M | 3 | 3.92 GB| very small, high quality loss |
| [Qwen1.5-7B-Chat-Q3_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q3_K_S.gguf) | Q3_K_S | 3 | 3.57 GB| very small, high quality loss |
| [Qwen1.5-7B-Chat-Q4_0.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q4_0.gguf) | Q4_0 | 4 | 4.51 GB| legacy; small, very high quality loss - prefer using Q3_K_M |
| [Qwen1.5-7B-Chat-Q4_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q4_K_M.gguf) | Q4_K_M | 4 | 4.77 GB| medium, balanced quality - recommended |
| [Qwen1.5-7B-Chat-Q4_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q4_K_S.gguf) | Q4_K_S | 4 | 4.54 GB| small, greater quality loss |
| [Qwen1.5-7B-Chat-Q5_0.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q5_0.gguf) | Q5_0 | 5 | 5.40 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
| [Qwen1.5-7B-Chat-Q5_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q5_K_M.gguf) | Q5_K_M | 5 | 5.53 GB| large, very low quality loss - recommended |
| [Qwen1.5-7B-Chat-Q5_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q5_K_S.gguf) | Q5_K_S | 5 | 5.4 GB| large, low quality loss - recommended |
| [Qwen1.5-7B-Chat-Q6_K.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q6_K.gguf) | Q6_K | 6 | 6.34 GB| very large, extremely low quality loss |
| [Qwen1.5-7B-Chat-Q8_0.gguf](https://huggingface.co/second-state/Qwen1.5-7B-Chat-GGUF/blob/main/Qwen1.5-7B-Chat-Q8_0.gguf) | Q8_0 | 8 | 8.21 GB| very large, extremely low quality loss - not recommended |
| {"language": ["en"], "license": "other", "tags": ["chat"], "model_name": "Qwen1.5 7B Chat", "base_model": "Qwen/Qwen1.5-7B-Chat", "license_name": "tongyi-qianwen-research", "license_link": "https://huggingface.co/Qwen/Qwen1.5-7B-Chat/blob/main/LICENSE", "model_creator": "Qwen", "quantized_by": "Second State Inc.", "pipeline_tag": "text-generation"} | text-generation | second-state/Qwen1.5-7B-Chat-GGUF | [
"gguf",
"chat",
"text-generation",
"en",
"base_model:Qwen/Qwen1.5-7B-Chat",
"license:other",
"region:us"
] | 2024-02-06T06:59:47+00:00 | [] | [
"en"
] | TAGS
#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-7B-Chat #license-other #region-us
|

---
Qwen1.5-7B-Chat-GGUF
====================
Original Model
--------------
Qwen/Qwen1.5-7B-Chat
Run with LlamaEdge
------------------
* LlamaEdge version: v0.2.15 and above
* Prompt template
+ Prompt type: 'chatml'
+ Prompt string
* Run as LlamaEdge service
* Run as LlamaEdge command app
Quantized GGUF Models
---------------------
| [] | [
"TAGS\n#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-7B-Chat #license-other #region-us \n"
] | [
38
] | [
"passage: TAGS\n#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-7B-Chat #license-other #region-us \n"
] | [
-0.010023923590779305,
-0.012251344509422779,
-0.003757887752726674,
-0.015251122415065765,
0.007713848724961281,
0.004232623614370823,
0.19796787202358246,
0.08698110282421112,
0.20624151825904846,
-0.05518682673573494,
0.0983501598238945,
-0.026160750538110733,
0.05950681492686272,
0.09443717449903488,
0.039224788546562195,
-0.17634575068950653,
0.06330893933773041,
-0.04392767697572708,
-0.006724408827722073,
0.058646589517593384,
0.06345642358064651,
0.01222515944391489,
0.08729035407304764,
-0.001578928087837994,
-0.08942367136478424,
0.034182898700237274,
-0.017799755558371544,
-0.029970820993185043,
0.0947447195649147,
0.10802784562110901,
-0.012084118090569973,
0.10158795118331909,
-0.047085728496313095,
-0.2238788604736328,
0.040655504912137985,
-0.055112436413764954,
-0.14940054714679718,
-0.001322337193414569,
0.0022816488053649664,
0.026355119422078133,
0.08851133286952972,
0.1978902369737625,
-0.08572524040937424,
0.12045920640230179,
-0.14707957208156586,
-0.07667047530412674,
-0.0796353816986084,
0.09249403327703476,
0.028649281710386276,
0.053178127855062485,
-0.023667369037866592,
0.053584545850753784,
-0.1270279586315155,
0.0295062568038702,
0.09982670098543167,
-0.47993531823158264,
0.05782134085893631,
0.2783401608467102,
0.11605481803417206,
0.13088572025299072,
-0.0846443697810173,
0.04907289519906044,
0.054699648171663284,
-0.020448187366127968,
-0.18679510056972504,
-0.026139812543988228,
0.07000195980072021,
0.07371638715267181,
-0.05420982837677002,
-0.016504142433404922,
0.24165284633636475,
0.07546748220920563,
-0.007682814262807369,
0.026165025308728218,
-0.01721412129700184,
-0.01232738047838211,
-0.02539721503853798,
0.05879060551524162,
-0.04141927510499954,
0.17617562413215637,
0.01376855093985796,
-0.08587073534727097,
-0.10659762471914291,
-0.05330963805317879,
-0.17971813678741455,
0.004171591252088547,
0.0007411452243104577,
0.08800608664751053,
-0.18538491427898407,
0.055729422718286514,
-0.1516152173280716,
-0.0751345306634903,
-0.08235028386116028,
-0.11885504424571991,
0.022777119651436806,
0.06275609880685806,
0.007594841066747904,
0.12458325922489166,
0.20343081653118134,
0.15383240580558777,
-0.0754973292350769,
0.07426387071609497,
-0.06962531059980392,
0.11244838684797287,
0.03079424798488617,
0.0014324623625725508,
0.04938611388206482,
-0.020431602373719215,
0.04493137449026108,
-0.10394547879695892,
-0.024630989879369736,
-0.04644649475812912,
-0.1667945683002472,
0.001871955581009388,
-0.12263824045658112,
0.08488496392965317,
0.008875164203345776,
0.02312012016773224,
0.009378190152347088,
-0.0013171377358958125,
0.23048198223114014,
0.04705561324954033,
-0.06197035312652588,
0.05247525870800018,
-0.007846425287425518,
-0.10682621598243713,
-0.02418529987335205,
0.10156308114528656,
0.07020393759012222,
-0.13015401363372803,
-0.06304170191287994,
-0.08907349407672882,
0.040130775421857834,
0.03674157336354256,
0.03836400806903839,
0.04959690570831299,
0.024255970492959023,
-0.13698351383209229,
-0.1040869653224945,
0.03234555199742317,
-0.00008219444862334058,
-0.03035220131278038,
-0.072097547352314,
-0.047435082495212555,
-0.042849138379096985,
-0.005659756250679493,
-0.05793085694313049,
-0.005929409526288509,
-0.09039118140935898,
0.07404056936502457,
-0.04202018305659294,
0.006034242454916239,
-0.2606058716773987,
-0.0018217499600723386,
-0.03930887207388878,
0.028005702421069145,
-0.07353438436985016,
0.0884525403380394,
-0.175022155046463,
0.10268606245517731,
-0.014943057671189308,
0.05117874592542648,
-0.09344393014907837,
0.014684979803860188,
-0.07018837332725525,
0.16153791546821594,
-0.10858219861984253,
-0.09106181561946869,
0.2716147005558014,
-0.10057516396045685,
-0.13298547267913818,
0.08447109162807465,
-0.003280648961663246,
-0.01404445618391037,
0.050753623247146606,
0.40039920806884766,
0.046420592814683914,
-0.05837889015674591,
-0.01797918975353241,
0.17493514716625214,
-0.11369805037975311,
-0.06856326013803482,
0.08680742979049683,
-0.08076582849025726,
-0.08806788176298141,
0.038647208362817764,
0.021513154730200768,
0.13540054857730865,
-0.05078567937016487,
-0.06729613244533539,
-0.01752576418220997,
-0.04936171695590019,
0.020689789205789566,
-0.030006539076566696,
0.06914723664522171,
-0.05636425316333771,
-0.007068020757287741,
-0.10804732888936996,
0.07533719390630722,
0.08213639259338379,
-0.027974599972367287,
-0.13122999668121338,
0.0611213855445385,
0.10673598200082779,
0.04982884228229523,
0.001932550803758204,
-0.09702152013778687,
0.010968332178890705,
0.05374796688556671,
0.09290917962789536,
0.11979266256093979,
0.05658698081970215,
-0.03043409436941147,
0.0028740877751260996,
-0.008192047476768494,
0.03932782635092735,
-0.01148014236241579,
-0.038476310670375824,
-0.1029018685221672,
0.0329175740480423,
-0.044246021658182144,
0.11861501634120941,
-0.11010003089904785,
0.00009592232527211308,
0.1019970029592514,
0.048945315182209015,
-0.054537124931812286,
0.03725263476371765,
-0.014149720780551434,
-0.03981434553861618,
0.006781395059078932,
-0.004989818204194307,
0.0733717530965805,
0.027323678135871887,
-0.12413661926984787,
0.22748209536075592,
-0.05176663398742676,
0.09270085394382477,
0.11746545135974884,
-0.014872683212161064,
0.06966222822666168,
-0.12980917096138,
-0.06082531809806824,
-0.005913549102842808,
0.11399701982736588,
-0.02574768289923668,
0.21857130527496338,
-0.04957639425992966,
0.07484867423772812,
-0.07230167090892792,
-0.05126657336950302,
-0.018908236175775528,
0.007123521529138088,
-0.024705534800887108,
0.10319137573242188,
0.08758977800607681,
-0.2036685049533844,
0.10801944136619568,
0.17795208096504211,
0.23986698687076569,
0.2762722074985504,
-0.08745583146810532,
0.05340023338794708,
-0.01472392212599516,
0.06501207500696182,
-0.08516129851341248,
0.12451983988285065,
-0.21665295958518982,
-0.01402337197214365,
0.020286327227950096,
0.05594233050942421,
0.13192231953144073,
-0.0649246945977211,
-0.12091577798128128,
-0.04807071015238762,
-0.09890205413103104,
-0.14358918368816376,
0.10451041907072067,
-0.040552981197834015,
0.05537585914134979,
0.021438319236040115,
-0.030793337151408195,
0.08889581263065338,
-0.03959157317876816,
-0.075214684009552,
0.14149188995361328,
-0.1491786241531372,
-0.1698676347732544,
-0.04697435721755028,
-0.06817153841257095,
-0.08314086496829987,
-0.0013930127024650574,
0.10013355314731598,
-0.1671188622713089,
0.012071016244590282,
0.0060391915030777454,
-0.03286195918917656,
-0.09974377602338791,
-0.0020467594731599092,
0.015684282407164574,
-0.016994722187519073,
-0.04856031388044357,
-0.15056253969669342,
-0.03760746121406555,
-0.0413692370057106,
-0.1162550300359726,
0.08740314841270447,
-0.1636502593755722,
0.01188002061098814,
0.125803604722023,
0.05644999071955681,
0.08531446754932404,
-0.052023597061634064,
0.3481449484825134,
-0.06268011778593063,
-0.007911842316389084,
0.06429421156644821,
0.07858460396528244,
0.05948299542069435,
0.13597263395786285,
0.027900593355298042,
-0.10066819190979004,
0.0011199056170880795,
0.0135554950684309,
-0.07101839035749435,
-0.18056470155715942,
-0.11131304502487183,
-0.03577788546681404,
0.16412869095802307,
-0.10046332329511642,
0.14978621900081635,
0.12959809601306915,
-0.02185419388115406,
0.00960211269557476,
-0.07414934784173965,
0.060512032359838486,
0.0349406935274601,
0.03770527243614197,
-0.0549842044711113,
0.05795174464583397,
-0.06386516243219376,
0.05958009511232376,
0.1313602775335312,
0.11337856948375702,
0.09635602682828903,
0.1532210111618042,
0.07292932271957397,
0.11878902465105057,
0.2647783160209656,
0.18746697902679443,
-0.14004696905612946,
-0.01745370402932167,
-0.09093192964792252,
-0.03156008571386337,
-0.02638440951704979,
0.028681108728051186,
0.011142930947244167,
-0.0026661199517548084,
-0.19842851161956787,
-0.02123044803738594,
-0.09620079398155212,
0.03712379187345505,
-0.009423254057765007,
0.0342300608754158,
0.08665638417005539,
0.003478112630546093,
0.048700593411922455,
-0.023129746317863464,
-0.007960282266139984,
0.13475218415260315,
0.06719771772623062,
-0.11169145256280899,
0.020901039242744446,
0.03867223858833313,
0.10539843142032623,
0.0625506341457367,
0.05565671622753143,
-0.09108513593673706,
-0.08701397478580475,
0.004153893794864416,
0.1199096068739891,
-0.19362081587314606,
0.2254386693239212,
0.005390425678342581,
-0.051606398075819016,
-0.03951449692249298,
-0.07837982475757599,
0.005617121234536171,
0.12987037003040314,
0.09910975396633148,
0.0704583004117012,
-0.09048110246658325,
0.013281665742397308,
-0.04846341162919998,
0.09901653975248337,
0.07341813296079636,
-0.032022833824157715,
-0.09208746999502182,
0.0035440644714981318,
0.023632677271962166,
-0.006245782133191824,
0.061292290687561035,
-0.14773714542388916,
-0.10130587220191956,
-0.01758076250553131,
0.07707910984754562,
0.11236275732517242,
-0.05559220910072327,
0.06512809544801712,
-0.08181210607290268,
0.08516019582748413,
-0.082713283598423,
-0.05478212982416153,
-0.050590820610523224,
-0.0788302943110466,
0.014886092394590378,
-0.05661996081471443,
-0.020255466923117638,
-0.05374392494559288,
-0.026733197271823883,
-0.10454223304986954,
-0.21617881953716278,
0.07812090218067169,
-0.07627750933170319,
-0.049150336533784866,
-0.014093101024627686,
0.1996528059244156,
0.0465274341404438,
0.0026245901826769114,
0.04502534866333008,
-0.08964557945728302,
-0.06805925071239471,
-0.18798555433750153,
0.05190025642514229,
-0.11218465864658356,
-0.11701711267232895,
-0.001820153440348804,
0.020756613463163376,
0.008447119034826756,
-0.025532668456435204,
-0.06773488968610764,
0.16796275973320007,
0.3197345733642578,
-0.050913140177726746,
0.22266581654548645,
0.2744908928871155,
-0.059905875474214554,
-0.13883690536022186,
-0.19411368668079376,
-0.10639799386262894,
-0.07417057454586029,
0.031147882342338562,
-0.17043839395046234,
-0.015235736034810543,
0.06756298989057541,
-0.05304459109902382,
0.14239895343780518,
-0.2789377272129059,
-0.03581896424293518,
0.09469030052423477,
-0.08148606866598129,
0.41732892394065857,
-0.13207611441612244,
-0.171656996011734,
0.00730784609913826,
-0.2561968266963959,
0.13962802290916443,
-0.13413818180561066,
0.09269554167985916,
0.015904275700449944,
0.07839807122945786,
-0.035936322063207626,
0.0076913065277040005,
0.13928477466106415,
0.04554441571235657,
0.017151908949017525,
-0.08913202583789825,
-0.06572521477937698,
0.08024165779352188,
0.046336084604263306,
-0.0363786518573761,
-0.1709577739238739,
0.05155675858259201,
-0.156321719288826,
0.013511708937585354,
-0.12427856028079987,
0.03657197579741478,
-0.01021038182079792,
-0.06395906955003738,
-0.09157150983810425,
0.03185390308499336,
-0.056417737156152725,
0.033868882805109024,
0.12795060873031616,
-0.08355280756950378,
0.13655392825603485,
0.1220160648226738,
-0.03098439611494541,
-0.20394185185432434,
0.011755036190152168,
-0.10714548081159592,
-0.05155395716428757,
0.07165262848138809,
-0.17700079083442688,
-0.03432949259877205,
0.09166853874921799,
-0.02497405745089054,
0.08123353868722916,
0.05204411968588829,
-0.10881523787975311,
0.11489733308553696,
0.12699677050113678,
-0.14041046798229218,
-0.21957515180110931,
-0.06637277454137802,
0.0953708365559578,
0.17980149388313293,
0.07957257330417633,
0.0945713222026825,
0.0327179990708828,
-0.01816440187394619,
0.02222854271531105,
-0.012798311188817024,
-0.11602113395929337,
-0.057095639407634735,
0.04720175266265869,
0.012007436715066433,
-0.15058647096157074,
0.10207520425319672,
0.021977849304676056,
0.10818006843328476,
0.014859432354569435,
0.16622567176818848,
-0.13278454542160034,
-0.08868588507175446,
-0.19394618272781372,
0.15229344367980957,
-0.1513862907886505,
-0.029643742367625237,
0.013532040640711784,
-0.07638934999704361,
-0.0016297498950734735,
0.1312560737133026,
-0.027802808210253716,
0.09867507964372635,
0.09267042577266693,
0.015088045969605446,
0.11901017278432846,
-0.05232375115156174,
-0.14731357991695404,
-0.009579887613654137,
-0.08762480318546295,
-0.08813775330781937,
-0.00618865666911006,
0.15171925723552704,
-0.04086358845233917,
-0.12411273270845413,
-0.20248830318450928,
0.0075748334638774395,
-0.1629529744386673,
-0.05604002997279167,
-0.07359658181667328,
-0.012533684261143208,
-0.013969478197395802,
-0.1020304262638092,
-0.021003559231758118,
-0.0053457459434866905,
-0.10947755724191666,
-0.016052357852458954,
0.032969702035188675,
0.07378703355789185,
-0.09595706313848495,
0.026083827018737793,
0.08539731055498123,
0.07534655928611755,
0.16897866129875183,
0.12349878251552582,
-0.008260453119874,
0.1447283923625946,
-0.16399097442626953,
-0.03155883029103279,
0.00621916726231575,
-0.0015224760863929987,
0.03615089878439903,
0.1553276628255844,
-0.044248513877391815,
-0.014109388925135136,
-0.025613227859139442,
0.056519802659749985,
0.017789164558053017,
-0.12448640912771225,
-0.09543382376432419,
-0.016902968287467957,
-0.042601775377988815,
-0.015049964189529419,
-0.072492316365242,
0.12645792961120605,
0.04817895218729973,
0.05000751465559006,
-0.048319123685359955,
-0.0014162982115522027,
-0.07783212512731552,
0.0007021420169621706,
0.04237077757716179,
-0.048248302191495895,
0.016653604805469513,
-0.08352819830179214,
-0.07525946944952011,
-0.025059381499886513,
0.2293080985546112,
-0.06287047266960144,
-0.07421161979436874,
0.02735052816569805,
0.012533596716821194,
0.04819774627685547,
-0.037859804928302765,
0.2834360599517822,
0.08016263693571091,
-0.000752401479985565,
-0.05049600452184677,
0.055444423109292984,
-0.043959539383649826,
-0.16099867224693298,
0.02111901342868805,
-0.0019346146145835519,
0.018223382532596588,
0.004652897361665964,
0.18914444744586945,
-0.06758378446102142,
0.006365519016981125,
-0.053515661507844925,
-0.0028240052051842213,
-0.021158607676625252,
-0.08315559476613998,
0.0571269765496254,
0.19044925272464752,
-0.03668301925063133,
0.03923413157463074,
0.007066353224217892,
-0.028919514268636703,
-0.0614197663962841,
-0.1012532114982605,
-0.025549450889229774,
-0.21395628154277802,
0.08371850103139877,
-0.08372105658054352,
0.07395737618207932,
0.0845680832862854,
0.053783245384693146,
-0.05379843711853027,
-0.013822183944284916,
-0.010274802334606647,
-0.10424934327602386,
0.017424123361706734,
-0.04542282223701477,
0.006077682599425316,
-0.09518029540777206,
-0.001276101334951818,
0.0467776395380497,
-0.01605568267405033,
-0.1045437678694725,
0.0702613964676857,
-0.010431397706270218,
-0.0007344945333898067,
-0.09809619188308716,
-0.043190307915210724,
-0.044419631361961365,
0.06074334308505058,
0.013832615688443184,
0.19537028670310974,
0.021023375913500786,
0.009860659018158913,
0.046858854591846466,
0.2153526097536087,
0.02382626198232174,
-0.09063856303691864,
-0.02387833222746849,
-0.0297419261187315,
-0.0666034072637558,
0.10145694017410278,
-0.07332275807857513,
-0.002238942077383399,
-0.06767140328884125,
0.2671719491481781,
0.24685664474964142,
-0.1094135195016861,
0.02535463497042656,
-0.03213929384946823,
0.021514086052775383,
0.0726078450679779,
0.11276447027921677,
0.06158878654241562,
0.25239625573158264,
-0.10134154558181763,
-0.013402404263615608,
0.014223432168364525,
0.02502511255443096,
-0.09824591875076294,
0.0988076776266098,
0.021551424637436867,
-0.07005030661821365,
-0.05816686898469925,
0.09795695543289185,
-0.27659377455711365,
0.08965849876403809,
-0.14744529128074646,
-0.11713728308677673,
-0.023291196674108505,
-0.08688579499721527,
0.08346022665500641,
0.08770754188299179,
0.05533168464899063,
-0.06371259689331055,
-0.0016093795420601964,
0.023152658715844154,
-0.0009818332036957145,
-0.2426009476184845,
-0.1508997678756714,
0.08139016479253769,
0.018214290961623192,
0.11880751699209213,
-0.027569910511374474,
0.04634078964591026,
0.05594157427549362,
0.03259173408150673,
-0.05303821712732315,
0.033424027264118195,
0.021286526694893837,
-0.030276693403720856,
-0.1479313224554062,
-0.08346661925315857,
0.06849414110183716,
-0.005837393458932638,
0.04319389536976814,
-0.05600105598568916,
0.041222091764211655,
0.125286266207695,
-0.008098647929728031,
-0.0640704482793808,
0.07995270192623138,
-0.08157171308994293,
0.09252039343118668,
-0.011062311939895153,
-0.030327044427394867,
-0.05281958729028702,
-0.010574379935860634,
0.027135133743286133,
0.022896168753504753,
-0.055297642946243286,
-0.024842960759997368,
-0.04124285653233528,
-0.03988371416926384,
0.19386881589889526,
-0.02927788719534874,
-0.1509629786014557,
0.0008124879677779973,
-0.09296802431344986,
0.11784172058105469,
-0.013414856977760792,
0.13413509726524353,
0.2003892958164215,
0.022827358916401863,
0.016896627843379974,
-0.059467870742082596,
0.035245515406131744,
0.0063604870811104774,
-0.04214492812752724,
-0.07924312353134155
] |
null | null | null | https://civitai.com/models/122359/detail-tweaker-xl?modelVersionId=135867 | {} | null | Yuriy81/DetailTweakerXL_lora | [
"region:us"
] | 2024-02-06T07:01:14+00:00 | [] | [] | TAGS
#region-us
| URL | [] | [
"TAGS\n#region-us \n"
] | [
6
] | [
"passage: TAGS\n#region-us \n"
] | [
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my_awesome_eli5_clm-model
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 3.7407
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 3.9157 | 1.0 | 560 | 3.7574 |
| 3.8162 | 2.0 | 1120 | 3.7432 |
| 3.7746 | 3.0 | 1680 | 3.7407 |
### Framework versions
- Transformers 4.37.0
- Pytorch 2.1.2
- Datasets 2.1.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "distilgpt2", "model-index": [{"name": "my_awesome_eli5_clm-model", "results": []}]} | text-generation | UjjwalP/my_awesome_eli5_clm-model | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:distilgpt2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T07:04:58+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-distilgpt2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| my\_awesome\_eli5\_clm-model
============================
This model is a fine-tuned version of distilgpt2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 3.7407
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Training results
### Framework versions
* Transformers 4.37.0
* Pytorch 2.1.2
* Datasets 2.1.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-distilgpt2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1"
] | [
77,
98,
4,
30
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-distilgpt2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1"
] | [
-0.09235620498657227,
0.09283239394426346,
-0.0020225364714860916,
0.10447591543197632,
0.12615281343460083,
0.000539363652933389,
0.15626034140586853,
0.12509731948375702,
-0.06289782375097275,
0.03875778988003731,
0.1471046507358551,
0.12405938655138016,
0.029295656830072403,
0.13688847422599792,
-0.055826954543590546,
-0.2088497281074524,
0.02179415337741375,
0.027372518554329872,
-0.05079090967774391,
0.12472070753574371,
0.09375898540019989,
-0.10906699299812317,
0.10384329408407211,
-0.004454485606402159,
-0.1558813899755478,
-0.002868825802579522,
0.013076601549983025,
-0.05099330097436905,
0.1328216940164566,
0.038572896271944046,
0.08721525967121124,
0.030221182852983475,
0.06507385522127151,
-0.18481658399105072,
0.01030514482408762,
0.06117221340537071,
-0.0011064735008403659,
0.08276117593050003,
0.04699065536260605,
0.01198330894112587,
0.10392652451992035,
-0.0742967277765274,
0.0444403700530529,
0.026697315275669098,
-0.13013668358325958,
-0.2279496192932129,
-0.0810488685965538,
0.03052825666964054,
0.10487755388021469,
0.09541531652212143,
-0.017232177779078484,
0.13170857727527618,
-0.012169293127954006,
0.09681858867406845,
0.20439428091049194,
-0.3174065947532654,
-0.06433127820491791,
0.06165824085474014,
0.06146427243947983,
0.10484980046749115,
-0.0972668007016182,
-0.0033689679112285376,
0.05406650900840759,
0.035408865660429,
0.15016452968120575,
-0.03280862048268318,
0.0014558214461430907,
-0.0017718076705932617,
-0.14142288267612457,
-0.03940148651599884,
0.18265728652477264,
0.05362269654870033,
-0.05045206472277641,
-0.0744197890162468,
-0.0733284279704094,
-0.12629924714565277,
-0.0341125950217247,
-0.012663044966757298,
0.04335355386137962,
-0.006348961964249611,
-0.058717288076877594,
-0.05940502509474754,
-0.11233486235141754,
-0.0745992362499237,
-0.0285712331533432,
0.11793992668390274,
0.02458838000893593,
0.001568958512507379,
-0.01729440502822399,
0.1043943241238594,
-0.026323389261960983,
-0.14170266687870026,
0.002318240702152252,
0.021827781572937965,
0.02736584283411503,
-0.03359480947256088,
-0.04628951475024223,
-0.09633547067642212,
0.028907280415296555,
0.1272866129875183,
-0.05826945975422859,
0.058992497622966766,
0.0026333117857575417,
0.04868187755346298,
-0.10589595139026642,
0.1777971386909485,
-0.0318218357861042,
-0.045279182493686676,
0.02807564102113247,
0.09106813371181488,
0.07216423749923706,
-0.017960399389266968,
-0.1422339528799057,
0.031706489622592926,
0.11480294167995453,
0.025003742426633835,
-0.0374629981815815,
0.07015909254550934,
-0.04886306822299957,
-0.015091143548488617,
0.05042630806565285,
-0.0771135464310646,
0.00922354869544506,
-0.018605612218379974,
-0.041303906589746475,
-0.08431626856327057,
0.024422062560915947,
0.029643220826983452,
-0.0019999388605356216,
0.074957475066185,
-0.08580932021141052,
-0.0018625764641910791,
-0.07208416610956192,
-0.1068725660443306,
0.01645689643919468,
-0.07124849408864975,
0.021063804626464844,
-0.11258570849895477,
-0.20908363163471222,
-0.0001961410162039101,
0.058937184512615204,
-0.03487224131822586,
-0.05545150116086006,
-0.05601894110441208,
-0.07555311918258667,
0.01818564534187317,
-0.018923774361610413,
0.09616906195878983,
-0.07157089561223984,
0.10120551288127899,
0.048932306468486786,
0.056291382759809494,
-0.06327149271965027,
0.031547851860523224,
-0.10499998182058334,
0.04691363126039505,
-0.16234497725963593,
0.03242069110274315,
-0.04431934282183647,
0.07468489557504654,
-0.09510426968336105,
-0.06907248497009277,
-0.0143553726375103,
-0.005951075814664364,
0.06510437279939651,
0.09876417368650436,
-0.14795996248722076,
-0.06363547593355179,
0.17147478461265564,
-0.08518875390291214,
-0.16626830399036407,
0.14262862503528595,
-0.034873951226472855,
0.05618010088801384,
0.06875952333211899,
0.22000715136528015,
0.0584852360188961,
-0.09051454812288284,
0.011429106816649437,
0.008895604871213436,
0.06046994403004646,
-0.054589614272117615,
0.09824904054403305,
-0.01096261478960514,
0.009929870255291462,
0.015875600278377533,
-0.0592464841902256,
0.06632261723279953,
-0.060303620994091034,
-0.08045709133148193,
-0.0441245473921299,
-0.10806496441364288,
0.03959846869111061,
0.04304203763604164,
0.06011442840099335,
-0.11878576874732971,
-0.09976773709058762,
0.031198663637042046,
0.07294692099094391,
-0.07892999798059464,
0.024586349725723267,
-0.07123718410730362,
0.10002132505178452,
-0.0938652977347374,
-0.0034179934300482273,
-0.13256649672985077,
-0.04283736273646355,
0.01884550414979458,
0.007938717491924763,
0.009419873356819153,
-0.005978844128549099,
0.07953236997127533,
0.08319421112537384,
-0.07303857803344727,
-0.04059198126196861,
-0.00828841794282198,
0.011137142777442932,
-0.12279845774173737,
-0.17745035886764526,
-0.012540731579065323,
-0.02897554263472557,
0.1539025902748108,
-0.2233370989561081,
0.055781446397304535,
0.005093411076813936,
0.0710325837135315,
0.029421159997582436,
-0.026312094181776047,
-0.02846500463783741,
0.038885921239852905,
-0.048779718577861786,
-0.07960096001625061,
0.0690583661198616,
0.029874639585614204,
-0.11327370256185532,
-0.004308327566832304,
-0.17398229241371155,
0.18866118788719177,
0.13291753828525543,
-0.055064596235752106,
-0.05490332096815109,
0.008937040343880653,
-0.03875577822327614,
-0.029686342924833298,
-0.03103865124285221,
-0.017028992995619774,
0.12723441421985626,
-0.004303293768316507,
0.16419678926467896,
-0.09754981845617294,
-0.051460929214954376,
0.03390691056847572,
-0.047758400440216064,
0.0049142008647322655,
0.11114531755447388,
0.060106899589300156,
-0.12600383162498474,
0.16190959513187408,
0.1990491896867752,
-0.057746414095163345,
0.14138035476207733,
-0.03745795041322708,
-0.058402616530656815,
-0.029853036627173424,
0.02635631524026394,
0.015546556562185287,
0.1133822500705719,
-0.11634399741888046,
0.00938242208212614,
0.011017773300409317,
0.026778114959597588,
0.01671660505235195,
-0.21128159761428833,
-0.031373947858810425,
0.04562284052371979,
-0.06029845029115677,
0.007340754382312298,
-0.011359892785549164,
-0.03560424596071243,
0.08798366039991379,
0.009242009371519089,
-0.06270745396614075,
0.05055234581232071,
0.0017524359282106161,
-0.08228850364685059,
0.20319445431232452,
-0.07232557982206345,
-0.1804957389831543,
-0.14534735679626465,
-0.07329137623310089,
-0.06525875627994537,
0.042244479060173035,
0.07573089748620987,
-0.0684470385313034,
-0.04339185729622841,
-0.12505225837230682,
0.0046225013211369514,
-0.0027336396742612123,
0.014274406246840954,
0.024423791095614433,
-0.012450449168682098,
0.08542145788669586,
-0.09760031849145889,
-0.018883362412452698,
-0.014558392576873302,
-0.04256833717226982,
0.026563171297311783,
0.01609335094690323,
0.11269807815551758,
0.12825262546539307,
-0.0016830541426315904,
0.002540862886235118,
-0.03162936493754387,
0.2464044988155365,
-0.07466480135917664,
-0.006606076378375292,
0.15195490419864655,
-0.013336388394236565,
0.05766922980546951,
0.1219402402639389,
0.042751699686050415,
-0.10523246228694916,
0.018917283043265343,
0.021958449855446815,
-0.041951872408390045,
-0.19590066373348236,
-0.024578096345067024,
-0.0473310686647892,
0.02483656443655491,
0.09134228527545929,
0.04257062450051308,
0.05034105107188225,
0.06656848639249802,
0.007088610902428627,
0.08499427884817123,
0.006133240647614002,
0.0869666039943695,
0.11598101258277893,
0.03667883202433586,
0.12956362962722778,
-0.051863424479961395,
-0.04309951141476631,
0.043728090822696686,
0.017970891669392586,
0.18414773046970367,
0.021787777543067932,
0.17321206629276276,
0.0438387393951416,
0.1491815149784088,
-0.0065379212610423565,
0.05360177159309387,
-0.017132258042693138,
-0.03969099000096321,
-0.023847969248890877,
-0.04743635654449463,
-0.03342963382601738,
0.0365430973470211,
-0.08430042862892151,
0.049903761595487595,
-0.0914410874247551,
0.007796091493219137,
0.05676140636205673,
0.23891346156597137,
0.05091305077075958,
-0.33781343698501587,
-0.1016736775636673,
0.040425170212984085,
-0.017840219661593437,
-0.040847547352313995,
0.02354769967496395,
0.13521051406860352,
-0.06204742565751076,
0.04077530279755592,
-0.08483190089464188,
0.084236279129982,
-0.03705611452460289,
0.05171556770801544,
0.05007369443774223,
0.07183549553155899,
-0.0082771684974432,
0.07762432843446732,
-0.2808650732040405,
0.26079660654067993,
0.017317334190011024,
0.07057910412549973,
-0.054769158363342285,
0.008123253472149372,
0.02424859255552292,
0.08073517680168152,
0.09457074105739594,
-0.011472362093627453,
-0.059547100216150284,
-0.18504250049591064,
-0.07271970063447952,
0.02811620384454727,
0.08289452642202377,
-0.05312241241335869,
0.10055717080831528,
-0.04739145189523697,
0.005117824766784906,
0.07427450269460678,
-0.0012605608208104968,
-0.05528191849589348,
-0.10828781872987747,
0.005826995708048344,
0.047289807349443436,
-0.0159402284771204,
-0.08636731654405594,
-0.08138793706893921,
-0.11330419033765793,
0.17110833525657654,
-0.04276515915989876,
-0.05153600126504898,
-0.10007879137992859,
0.04635896906256676,
0.04589630290865898,
-0.08024710416793823,
0.03776237741112709,
0.0018978692824020982,
0.09611813724040985,
0.01918874680995941,
-0.061461497098207474,
0.11849083751440048,
-0.0759269967675209,
-0.16741199791431427,
-0.05068952962756157,
0.1287691444158554,
-0.0008354694582521915,
0.04547818750143051,
-0.001999868545681238,
0.021314168348908424,
-0.03842945396900177,
-0.07850256562232971,
0.033785510808229446,
-0.011240066960453987,
0.05974634736776352,
-0.02323378250002861,
-0.023847833275794983,
0.0344509482383728,
-0.059467755258083344,
-0.05527065694332123,
0.1636110097169876,
0.3016613721847534,
-0.07306861132383347,
0.02342204935848713,
0.053990405052900314,
-0.06402242183685303,
-0.18663519620895386,
0.008081690408289433,
0.015802284702658653,
-0.002850584452971816,
0.06107491999864578,
-0.14141717553138733,
0.0747695118188858,
0.09842342883348465,
-0.0341801792383194,
0.12229830771684647,
-0.3133319914340973,
-0.13815699517726898,
0.10367163270711899,
0.15326735377311707,
0.12264645844697952,
-0.1740807145833969,
-0.05313153564929962,
-0.04595266655087471,
-0.1213718131184578,
0.10307591408491135,
-0.12258422374725342,
0.11632706969976425,
-0.007210337556898594,
0.054651692509651184,
0.00316413352265954,
-0.055387552827596664,
0.14555785059928894,
-0.040693555027246475,
0.08572469651699066,
-0.07437678426504135,
0.028674831613898277,
0.06920190900564194,
-0.06421158462762833,
0.02793809026479721,
-0.1502693146467209,
0.034034498035907745,
-0.06341665983200073,
-0.03474556282162666,
-0.05322672426700592,
0.030674608424305916,
-0.030879594385623932,
-0.06629367172718048,
-0.034922342747449875,
0.0202472060918808,
0.044496722519397736,
-0.011608053930103779,
0.16452965140342712,
0.0061037782579660416,
0.15105746686458588,
0.13794384896755219,
0.08268826454877853,
-0.08465016633272171,
-0.03038746491074562,
-0.01729629933834076,
-0.036861222237348557,
0.04725464805960655,
-0.16137586534023285,
0.03500406816601753,
0.10826722532510757,
-0.0019320512656122446,
0.14432132244110107,
0.06675611436367035,
-0.03853161260485649,
0.017815109342336655,
0.06344094127416611,
-0.17958316206932068,
-0.1507701426744461,
-0.031102044507861137,
-0.0067572458647191525,
-0.11662019789218903,
0.055456649512052536,
0.13104760646820068,
-0.0695062056183815,
0.002623139414936304,
-0.0061055393889546394,
0.02602965198457241,
-0.037649307399988174,
0.17096783220767975,
0.04369938373565674,
0.04944673925638199,
-0.07747197896242142,
0.08698584884405136,
0.043627869337797165,
-0.06877902150154114,
0.02577352337539196,
0.04865865036845207,
-0.08882295340299606,
-0.05068628117442131,
0.03142443299293518,
0.16392751038074493,
-0.03314767777919769,
-0.05397623032331467,
-0.14797542989253998,
-0.11133012920618057,
0.04892599955201149,
0.1372329443693161,
0.0899297371506691,
0.015432099811732769,
-0.02986122854053974,
0.014381918124854565,
-0.10783068835735321,
0.11847466975450516,
0.028004677966237068,
0.08536162227392197,
-0.16523078083992004,
0.09358424693346024,
0.0012002075091004372,
0.01796429604291916,
-0.021402806043624878,
0.040779776871204376,
-0.10049773752689362,
-0.003504143562167883,
-0.1324334293603897,
-0.010573695413768291,
-0.031302064657211304,
-0.0031639831140637398,
-0.008452082052826881,
-0.054541878402233124,
-0.05907640978693962,
0.02008720301091671,
-0.09736061096191406,
-0.03417490795254707,
0.025065883994102478,
0.04763234406709671,
-0.11713706701993942,
-0.03879721090197563,
0.02437761425971985,
-0.07378343492746353,
0.073297418653965,
0.014226438477635384,
0.018662771210074425,
0.0511133186519146,
-0.14814965426921844,
0.043409816920757294,
0.05534855276346207,
0.00647188862785697,
0.028434110805392265,
-0.08863189816474915,
-0.02426411397755146,
0.007483232766389847,
0.03669656440615654,
0.023426169529557228,
0.09375695884227753,
-0.12409432232379913,
0.0024514892138540745,
-0.019839219748973846,
-0.05399676412343979,
-0.05705447867512703,
0.031404752284288406,
0.065538190305233,
0.021386392414569855,
0.2090320736169815,
-0.09351019561290741,
0.009795812889933586,
-0.20426441729068756,
0.011399460025131702,
0.006745213642716408,
-0.12929248809814453,
-0.11020006239414215,
-0.05847710743546486,
0.044982679188251495,
-0.05486255884170532,
0.12795250117778778,
-0.0017312358831986785,
0.04679021239280701,
0.03876136988401413,
-0.014454858377575874,
0.07010937482118607,
0.010928839445114136,
0.2371138036251068,
0.015241077169775963,
-0.03930594399571419,
0.03580016642808914,
0.03659697249531746,
0.11320901662111282,
0.08266594260931015,
0.16053658723831177,
0.14796416461467743,
-0.04053681343793869,
0.10649517178535461,
0.02137972228229046,
-0.03667072206735611,
-0.16728642582893372,
0.04603351652622223,
-0.04043588042259216,
0.11231739073991776,
-0.016728535294532776,
0.2279050201177597,
0.11444055289030075,
-0.16308534145355225,
0.008809561841189861,
-0.05980643630027771,
-0.07310251891613007,
-0.1038016676902771,
-0.10079126805067062,
-0.09011264145374298,
-0.14889313280582428,
-0.007417869288474321,
-0.10823384672403336,
0.023327842354774475,
0.10085871815681458,
0.00633277790620923,
-0.0283975712954998,
0.17653459310531616,
0.013735873624682426,
0.008738442324101925,
0.03882066160440445,
-0.008738585747778416,
-0.039522506296634674,
-0.07431073486804962,
-0.08872313052415848,
0.00685076741501689,
-0.01024137157946825,
0.03565827012062073,
-0.040971431881189346,
-0.031912386417388916,
0.03646264970302582,
-0.021720070391893387,
-0.10608816891908646,
0.0031724772416055202,
0.03999015688896179,
0.05099628493189812,
0.0472140908241272,
0.008664604276418686,
-0.0023772979620844126,
0.0045419009402394295,
0.24418428540229797,
-0.07999027520418167,
-0.0645361840724945,
-0.0857553705573082,
0.20706012845039368,
0.007298321928828955,
-0.010074728168547153,
0.002647997811436653,
-0.08815022557973862,
0.03238634392619133,
0.23724447190761566,
0.17719022929668427,
-0.08630365133285522,
-0.009027091786265373,
-0.021606434136629105,
-0.005762002430856228,
-0.02816659025847912,
0.09561612457036972,
0.11391974985599518,
0.017039205878973007,
-0.08374236524105072,
-0.03060053661465645,
-0.03804957494139671,
-0.0015837022801861167,
-0.050607677549123764,
0.06736237555742264,
0.019124051555991173,
0.015856264159083366,
-0.03764011338353157,
0.06331446766853333,
-0.031141316518187523,
-0.08637687563896179,
0.001135742524638772,
-0.1883215755224228,
-0.13597345352172852,
-0.011007762514054775,
0.11390049010515213,
-0.02495625801384449,
0.04784010350704193,
-0.025448735803365707,
0.023208219558000565,
0.048318713903427124,
-0.024368589743971825,
-0.07101146131753922,
-0.05956929922103882,
0.06530626118183136,
-0.10675332695245743,
0.24045853316783905,
-0.0416053831577301,
0.024828119203448296,
0.12935014069080353,
0.029975702986121178,
-0.09783507883548737,
0.09681575000286102,
0.0431506521999836,
-0.043881796300411224,
0.041447028517723083,
0.0913805291056633,
-0.030077924951910973,
0.10413043946027756,
0.053110700100660324,
-0.11551517248153687,
0.004183911252766848,
-0.04414166510105133,
-0.061283912509679794,
-0.03530094772577286,
-0.048600804060697556,
-0.0664115846157074,
0.1411442905664444,
0.16137933731079102,
-0.04688812047243118,
-0.007113039959222078,
-0.05812336876988411,
0.022749774158000946,
0.07385729253292084,
0.03934509679675102,
-0.028096696361899376,
-0.24323999881744385,
0.00638643279671669,
0.07791154831647873,
-0.009963388554751873,
-0.30740559101104736,
-0.08721277862787247,
-0.013923629187047482,
-0.03459073603153229,
-0.09539259225130081,
0.08636302500963211,
0.12015801668167114,
0.038379985839128494,
-0.06506258994340897,
-0.06020032986998558,
-0.07784318178892136,
0.15846949815750122,
-0.12248724699020386,
-0.0953281819820404
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# checkpoints
This model is a fine-tuned version of [kavg/LiLT-RE-ZH](https://huggingface.co/kavg/LiLT-RE-ZH) on the xfun dataset.
It achieves the following results on the evaluation set:
- Precision: 0.4194
- Recall: 0.5166
- F1: 0.4629
- Loss: 0.1587
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 2
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 4000
### Training results
| Training Loss | Epoch | Step | F1 | Validation Loss | Precision | Recall |
|:-------------:|:------:|:----:|:------:|:---------------:|:---------:|:------:|
| 0.0851 | 41.67 | 250 | 0.3776 | 0.1733 | 0.4 | 0.3576 |
| 0.0596 | 83.33 | 500 | 0.4629 | 0.1587 | 0.4194 | 0.5166 |
| 0.0385 | 125.0 | 750 | 0.5181 | 0.2059 | 0.4471 | 0.6159 |
| 0.0198 | 166.67 | 1000 | 0.5311 | 0.2371 | 0.4631 | 0.6225 |
| 0.015 | 208.33 | 1250 | 0.5299 | 0.2241 | 0.465 | 0.6159 |
| 0.0169 | 250.0 | 1500 | 0.5057 | 0.2671 | 0.4467 | 0.5828 |
| 0.0158 | 291.67 | 1750 | 0.5341 | 0.2537 | 0.4839 | 0.5960 |
| 0.0184 | 333.33 | 2000 | 0.5187 | 0.2883 | 0.4592 | 0.5960 |
| 0.013 | 375.0 | 2250 | 0.5215 | 0.2755 | 0.4596 | 0.6026 |
| 0.0027 | 416.67 | 2500 | 0.5210 | 0.3146 | 0.4515 | 0.6159 |
| 0.0094 | 458.33 | 2750 | 0.5239 | 0.3298 | 0.4559 | 0.6159 |
| 0.0042 | 500.0 | 3000 | 0.5158 | 0.3348 | 0.4545 | 0.5960 |
| 0.0057 | 541.67 | 3250 | 0.5254 | 0.3423 | 0.4581 | 0.6159 |
| 0.0049 | 583.33 | 3500 | 0.5254 | 0.3517 | 0.4581 | 0.6159 |
| 0.0131 | 625.0 | 3750 | 0.5341 | 0.3328 | 0.4677 | 0.6225 |
| 0.0077 | 666.67 | 4000 | 0.5326 | 0.3384 | 0.4653 | 0.6225 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["xfun"], "metrics": ["precision", "recall", "f1"], "base_model": "kavg/LiLT-RE-ZH", "model-index": [{"name": "checkpoints", "results": []}]} | null | kavg/LiLT-RE-ZH-SIN | [
"transformers",
"safetensors",
"lilt",
"generated_from_trainer",
"dataset:xfun",
"base_model:kavg/LiLT-RE-ZH",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-06T07:05:30+00:00 | [] | [] | TAGS
#transformers #safetensors #lilt #generated_from_trainer #dataset-xfun #base_model-kavg/LiLT-RE-ZH #license-mit #endpoints_compatible #region-us
| checkpoints
===========
This model is a fine-tuned version of kavg/LiLT-RE-ZH on the xfun dataset.
It achieves the following results on the evaluation set:
* Precision: 0.4194
* Recall: 0.5166
* F1: 0.4629
* Loss: 0.1587
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 8
* eval\_batch\_size: 2
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 4000
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 4000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #lilt #generated_from_trainer #dataset-xfun #base_model-kavg/LiLT-RE-ZH #license-mit #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 4000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
57,
115,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #lilt #generated_from_trainer #dataset-xfun #base_model-kavg/LiLT-RE-ZH #license-mit #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 4000### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.1150847002863884,
0.03121502324938774,
-0.0011505711590871215,
0.09313024580478668,
0.1461421549320221,
0.018554477021098137,
0.10845281928777695,
0.11453615874052048,
-0.08778540790081024,
0.030916592106223106,
0.12830661237239838,
0.13907849788665771,
0.00880915205925703,
0.17564842104911804,
-0.059255219995975494,
-0.2351016253232956,
0.010735034942626953,
0.012398846447467804,
-0.07543341815471649,
0.11930808424949646,
0.09321895241737366,
-0.1539827138185501,
0.0870509147644043,
-0.0013411209220066667,
-0.18301448225975037,
-0.011668439954519272,
0.025547469034790993,
-0.03568201884627342,
0.1464405357837677,
0.015233189798891544,
0.12395405024290085,
0.042960140854120255,
0.11528582870960236,
-0.23075109720230103,
0.015976741909980774,
0.05407499149441719,
0.020789815112948418,
0.06585197150707245,
0.0307278111577034,
-0.013538154773414135,
0.10309188067913055,
-0.12818580865859985,
0.05418543890118599,
0.03046291135251522,
-0.13633805513381958,
-0.2570766508579254,
-0.11201946437358856,
0.036237869411706924,
0.08613121509552002,
0.07005342096090317,
-0.0033524376340210438,
0.12734244763851166,
-0.08157417923212051,
0.09218858927488327,
0.2667655646800995,
-0.3100087344646454,
-0.0639006569981575,
0.05739806219935417,
0.05096619203686714,
0.08373992145061493,
-0.13197048008441925,
-0.000035391280107432976,
0.056394871324300766,
0.008982090279459953,
0.16691751778125763,
-0.03430376946926117,
-0.06366107612848282,
0.029706666246056557,
-0.14872245490550995,
-0.02072959579527378,
0.0862649604678154,
0.0483633428812027,
-0.05387227237224579,
-0.044811177998781204,
-0.04833689704537392,
-0.15093453228473663,
-0.06831692159175873,
-0.013537430204451084,
0.05402011796832085,
-0.060445066541433334,
-0.08577201515436172,
-0.01185760647058487,
-0.10033216327428818,
-0.10136976093053818,
-0.015257025137543678,
0.20738807320594788,
0.05275311693549156,
0.0031771857757121325,
-0.015637122094631195,
0.12533606588840485,
-0.01340130902826786,
-0.14196498692035675,
-0.010961809195578098,
0.01531815342605114,
-0.036638155579566956,
-0.059733182191848755,
-0.04741816967725754,
-0.028457440435886383,
0.03304945304989815,
0.13321200013160706,
-0.10920634120702744,
0.048800937831401825,
0.036187391728162766,
0.018565881997346878,
-0.10376430302858353,
0.17617815732955933,
-0.06416739523410797,
-0.04480040445923805,
0.01501612737774849,
0.08524942398071289,
0.010351170785725117,
0.004822913557291031,
-0.0730270966887474,
0.0006549357785843313,
0.09421925246715546,
0.03709989786148071,
-0.09370321035385132,
0.07944975048303604,
-0.04218130558729172,
0.010705741122364998,
0.013038167729973793,
-0.08106981962919235,
0.016684168949723244,
0.009061995893716812,
-0.0786932110786438,
-0.0787949189543724,
0.018177349120378494,
0.020699597895145416,
0.027661586180329323,
0.10881713777780533,
-0.09668686240911484,
0.022987717762589455,
-0.08182547986507416,
-0.12217508256435394,
-0.00995084922760725,
-0.09025111049413681,
0.036730993539094925,
-0.10092448443174362,
-0.1642235368490219,
-0.017788046970963478,
0.05678369849920273,
-0.025632480159401894,
-0.0016786119667813182,
-0.05399216338992119,
-0.08647973090410233,
-0.0006542950286529958,
-0.005316458642482758,
0.11599444597959518,
-0.07549833506345749,
0.12068402022123337,
0.04047512635588646,
0.07684992253780365,
-0.011020353995263577,
0.026789622381329536,
-0.09167666733264923,
0.033352047204971313,
-0.20110087096691132,
0.024344108998775482,
-0.0822741910815239,
0.0727294459939003,
-0.09640593081712723,
-0.09142500162124634,
0.03928232565522194,
0.004574141930788755,
0.08869437128305435,
0.11103957891464233,
-0.15069882571697235,
-0.04720725491642952,
0.17870621383190155,
-0.08955162018537521,
-0.13662169873714447,
0.0976184532046318,
-0.06379331648349762,
0.07568296045064926,
0.07370108366012573,
0.21927623450756073,
0.08922414481639862,
-0.14198097586631775,
0.014198395423591137,
-0.018325187265872955,
0.028240269050002098,
-0.036429695785045624,
0.05358187481760979,
0.02339429222047329,
0.004948930349200964,
0.017164766788482666,
-0.08504056930541992,
0.063026063144207,
-0.11112304776906967,
-0.08263525366783142,
-0.03754410892724991,
-0.1293385624885559,
0.05707842484116554,
0.04120665788650513,
0.061365317553281784,
-0.11506030708551407,
-0.06690043210983276,
0.07095129787921906,
0.1072077825665474,
-0.06689159572124481,
0.02665233612060547,
-0.07742158323526382,
0.039665475487709045,
-0.03390561789274216,
-0.022862263023853302,
-0.14074651896953583,
-0.045583631843328476,
0.002529479330405593,
-0.010098443366587162,
-0.008695641532540321,
-0.008371826261281967,
0.08243963867425919,
0.07875458151102066,
-0.07132193446159363,
-0.046954210847616196,
-0.03875124081969261,
0.02467048354446888,
-0.11956454813480377,
-0.22705021500587463,
-0.03714332357048988,
-0.027512822300195694,
0.10862310975790024,
-0.23101435601711273,
0.03207908943295479,
-0.006370749324560165,
0.06488054990768433,
0.03442586585879326,
-0.009440772235393524,
-0.021680818870663643,
0.06255075335502625,
-0.0057894326746463776,
-0.06432231515645981,
0.05765119567513466,
0.0012604414951056242,
-0.07733527570962906,
-0.01036146655678749,
-0.10454168915748596,
0.16627736389636993,
0.10950002074241638,
-0.014754953794181347,
-0.09410183131694794,
-0.005784559063613415,
-0.055928785353899,
-0.03945863991975784,
-0.037098538130521774,
0.040592655539512634,
0.11451438814401627,
-0.003055092180147767,
0.13271024823188782,
-0.0773252323269844,
-0.0458366833627224,
0.03130796551704407,
-0.048656318336725235,
0.04324662312865257,
0.11603986471891403,
0.05222160741686821,
-0.11655793339014053,
0.11616355180740356,
0.13112741708755493,
-0.058036308735609055,
0.14799220860004425,
-0.05510048195719719,
-0.059664830565452576,
-0.03995557874441147,
0.015105699189007282,
0.0018278212519362569,
0.17260459065437317,
-0.0705600306391716,
0.0035126740112900734,
0.003033388638868928,
0.023613588884472847,
0.005180086009204388,
-0.22996951639652252,
-0.055286090821027756,
0.039211831986904144,
-0.06665048748254776,
-0.05090072751045227,
-0.0035640813875943422,
-0.01101740077137947,
0.0957089364528656,
0.0000015146395071496954,
-0.05845695734024048,
0.012262290343642235,
0.002817018423229456,
-0.06461916118860245,
0.22144095599651337,
-0.0792798325419426,
-0.06079662963747978,
-0.07237179577350616,
-0.05374602973461151,
-0.017860358580946922,
-0.007267139386385679,
0.06343396008014679,
-0.08175439387559891,
-0.01888473518192768,
-0.08060350269079208,
0.01702376827597618,
0.03320116549730301,
0.023350875824689865,
-0.013724514283239841,
-0.010058566927909851,
0.05207303911447525,
-0.1163836270570755,
-0.007032773457467556,
-0.06005256250500679,
-0.05857161059975624,
0.052478305995464325,
0.053223997354507446,
0.12956370413303375,
0.12563587725162506,
-0.02882622554898262,
0.007474309299141169,
-0.04461876302957535,
0.22096073627471924,
-0.07860594242811203,
-0.010220903903245926,
0.12228156626224518,
0.010925626382231712,
0.04750511795282364,
0.13503679633140564,
0.05354486405849457,
-0.10479225963354111,
0.01613612473011017,
0.034379392862319946,
-0.03348622843623161,
-0.20172138512134552,
-0.023849094286561012,
-0.03230677545070648,
-0.022724542766809464,
0.0632801279425621,
0.006770171225070953,
-0.002299556275829673,
0.06088896095752716,
0.02941550500690937,
0.017934728413820267,
-0.020146939903497696,
0.06139902025461197,
0.06509537249803543,
0.04109816253185272,
0.10871917754411697,
-0.04784172400832176,
-0.040846869349479675,
0.019518813118338585,
-0.002507081488147378,
0.2489246129989624,
-0.02611667849123478,
0.11590716242790222,
0.08195823431015015,
0.1854524463415146,
-0.013210676610469818,
0.06459124386310577,
0.0012377147795632482,
-0.06887675076723099,
-0.003176516154780984,
-0.04338470473885536,
-0.01582000032067299,
0.02996884472668171,
-0.08607529103755951,
0.04672056809067726,
-0.1332886666059494,
0.020079217851161957,
0.0641072615981102,
0.2621676027774811,
0.044285908341407776,
-0.3037569522857666,
-0.09636025875806808,
0.003727403236553073,
-0.016446556895971298,
-0.01907370239496231,
0.015038423240184784,
0.15097251534461975,
-0.08142055571079254,
0.030528662726283073,
-0.0753088966012001,
0.061438340693712234,
-0.011085571721196175,
0.03934727981686592,
0.049205232411623,
0.10743789374828339,
-0.03076520748436451,
0.050863511860370636,
-0.28303098678588867,
0.3128402531147003,
0.01200905255973339,
0.11086849123239517,
-0.043938811868429184,
-0.03464469313621521,
0.03111591562628746,
0.08858762681484222,
0.09622635692358017,
-0.006927056238055229,
-0.08179964870214462,
-0.2335987538099289,
-0.02420957013964653,
0.04704330861568451,
0.11216355860233307,
-0.026523591950535774,
0.1277616024017334,
-0.03351297602057457,
0.0013201824622228742,
0.07522035390138626,
-0.0253433957695961,
-0.100705087184906,
-0.052196204662323,
-0.034836724400520325,
0.0054759979248046875,
-0.004373325500637293,
-0.1016773208975792,
-0.08623571693897247,
-0.07936175912618637,
0.10043696314096451,
0.014340335503220558,
-0.031795576214790344,
-0.11666504293680191,
0.042618997395038605,
0.08089616894721985,
-0.08882971853017807,
0.049790870398283005,
0.03507349640130997,
0.055757418274879456,
0.016180608421564102,
-0.03353360295295715,
0.10948852449655533,
-0.07204730063676834,
-0.18157416582107544,
-0.049236636608839035,
0.10684192925691605,
0.055614814162254333,
0.04455922544002533,
0.005502636544406414,
0.009884585626423359,
-0.00023696207790635526,
-0.0915270522236824,
0.03497980162501335,
-0.02300262078642845,
0.0534684844315052,
-0.00573262432590127,
-0.05417684465646744,
0.03886789456009865,
-0.06204984337091446,
-0.04297567158937454,
0.15083380043506622,
0.31789037585258484,
-0.10252012312412262,
0.034385085105895996,
0.06445707380771637,
-0.052641790360212326,
-0.2050204575061798,
0.052422888576984406,
0.03971946984529495,
0.012820917181670666,
0.07118993997573853,
-0.15099161863327026,
0.07950292527675629,
0.0851702094078064,
-0.01823309063911438,
0.10172774642705917,
-0.33066099882125854,
-0.1279585212469101,
0.10588373243808746,
0.19091765582561493,
0.11505196243524551,
-0.15180137753486633,
-0.008235558867454529,
-0.0036909969057887793,
-0.12421398609876633,
0.08737976849079132,
-0.12673063576221466,
0.12913358211517334,
-0.025848297402262688,
0.05567131191492081,
0.018525056540966034,
-0.06489172577857971,
0.11025331914424896,
0.0072482586838305,
0.1355268508195877,
-0.04713789373636246,
-0.030294597148895264,
0.06832816451787949,
-0.07437041401863098,
0.02084505558013916,
-0.0865594670176506,
0.05471949651837349,
-0.04541440308094025,
-0.014154649339616299,
-0.09928345680236816,
0.030812101438641548,
-0.03865925222635269,
-0.07595788687467575,
-0.050908107310533524,
0.06684039533138275,
0.06343448907136917,
-0.02431684359908104,
0.12468760460615158,
0.006937957368791103,
0.13592734932899475,
0.08592189848423004,
0.044188883155584335,
-0.04787536710500717,
-0.02516014128923416,
0.0042896573431789875,
-0.018460461869835854,
0.051857445389032364,
-0.16574180126190186,
0.034210506826639175,
0.12451450526714325,
0.04848584160208702,
0.11931152641773224,
0.06120201200246811,
-0.04649181291460991,
0.012202948331832886,
0.05808033049106598,
-0.1458534598350525,
-0.11810684204101562,
0.0069115315563976765,
-0.07856852561235428,
-0.09162142872810364,
0.07568447291851044,
0.10972582548856735,
-0.08471882343292236,
0.001651128870435059,
-0.03840339183807373,
0.015543069690465927,
-0.0499303862452507,
0.2271275669336319,
0.08275990188121796,
0.06655991077423096,
-0.09184831380844116,
0.06556887179613113,
0.013704008422791958,
-0.024240225553512573,
0.009958350099623203,
0.03401155769824982,
-0.09844653308391571,
-0.035604119300842285,
0.0896587073802948,
0.15721555054187775,
-0.042080946266651154,
-0.035392604768276215,
-0.14426174759864807,
-0.12789496779441833,
0.042735498398542404,
0.2307918816804886,
0.08457636833190918,
0.021805118769407272,
-0.010136586613953114,
0.019436689093708992,
-0.13614216446876526,
0.11469139903783798,
0.03276699408888817,
0.0773310512304306,
-0.1540239006280899,
0.1787814497947693,
-0.017291739583015442,
0.04174404963850975,
-0.02263382077217102,
0.05189782753586769,
-0.14196033775806427,
0.010969803668558598,
-0.1571371853351593,
-0.027278240770101547,
-0.03345680609345436,
0.005016730632632971,
-0.013736656866967678,
-0.06845143437385559,
-0.06097888574004173,
0.03008795529603958,
-0.115464948117733,
-0.012093746103346348,
0.03838501498103142,
0.06200883910059929,
-0.1315070390701294,
-0.053012050688266754,
0.01642168127000332,
-0.06857538223266602,
0.05954062566161156,
0.02825670875608921,
0.0319589264690876,
0.06992999464273453,
-0.16592001914978027,
0.010658792220056057,
0.04196837171912193,
-0.024878906086087227,
0.06522686034440994,
-0.1193176656961441,
-0.02014082670211792,
-0.036206163465976715,
0.04846180975437164,
0.018212825059890747,
0.06002987176179886,
-0.13057005405426025,
0.011348077096045017,
-0.04121684655547142,
-0.03168096765875816,
-0.049814336001873016,
0.01731335185468197,
0.06763245165348053,
0.0129317007958889,
0.17685942351818085,
-0.0862746387720108,
0.019566643983125687,
-0.2104511857032776,
-0.023965559899806976,
-0.013733894564211369,
-0.11106424778699875,
-0.1151488795876503,
-0.054048940539360046,
0.06676086783409119,
-0.059894874691963196,
0.11465615779161453,
-0.010436142794787884,
0.042843297123909,
0.030347365885972977,
-0.06637175381183624,
0.026287401095032692,
0.02829192392528057,
0.2236524224281311,
-0.0014780426863580942,
-0.03386593982577324,
0.06656283885240555,
0.06159888580441475,
0.08350631594657898,
0.05887877196073532,
0.22005106508731842,
0.17379507422447205,
-0.012276443652808666,
0.08934139460325241,
0.029455408453941345,
-0.08538937568664551,
-0.13249745965003967,
0.05416293814778328,
-0.05599021166563034,
0.06230348348617554,
-0.03671171888709068,
0.20248554646968842,
0.11317484825849533,
-0.19014441967010498,
0.037418924272060394,
-0.04539221152663231,
-0.09264595061540604,
-0.09244120866060257,
-0.026322612538933754,
-0.08876994252204895,
-0.15871888399124146,
0.013107147067785263,
-0.1064019575715065,
0.026720112189650536,
0.10671480000019073,
0.0025501661002635956,
0.004029815085232258,
0.2076817899942398,
0.05548575147986412,
0.036874011158943176,
0.05343326926231384,
0.00479030329734087,
-0.05093837529420853,
-0.05433898791670799,
-0.07722925394773483,
0.023091498762369156,
-0.06122876703739166,
0.014860784634947777,
-0.06683643162250519,
-0.08176429569721222,
0.053111325949430466,
-0.003229300258681178,
-0.0949033796787262,
0.016098400577902794,
0.0363427959382534,
0.09757912158966064,
0.0103746447712183,
0.03564808517694473,
-0.0029356072191148996,
-0.00718914857134223,
0.266523540019989,
-0.05778510123491287,
-0.09049463272094727,
-0.09195166081190109,
0.26603084802627563,
0.07434094697237015,
0.0048876260407269,
0.030402107164263725,
-0.07760276645421982,
0.03971182182431221,
0.22074735164642334,
0.16506849229335785,
-0.08607702702283859,
0.012477889657020569,
-0.039674993604421616,
-0.014009841717779636,
-0.04162842035293579,
0.11537540704011917,
0.10326793789863586,
-0.00023128515749704093,
-0.0878826454281807,
-0.0416424460709095,
-0.05539781600236893,
-0.029776761308312416,
-0.027132796123623848,
0.06867191940546036,
0.04692262038588524,
0.007460957393050194,
-0.043792854994535446,
0.06601660698652267,
-0.04152552783489227,
-0.14397510886192322,
0.07029798626899719,
-0.18071413040161133,
-0.1402459591627121,
-0.04053577780723572,
0.09217000752687454,
0.000990956206806004,
0.06587035208940506,
-0.032612331211566925,
0.0028556084726005793,
0.06304792314767838,
-0.009701188653707504,
-0.04005136713385582,
-0.09953497350215912,
0.08517396450042725,
-0.06902443617582321,
0.2295438051223755,
-0.04776916652917862,
0.026572335511446,
0.12663881480693817,
0.029787685722112656,
-0.0943543016910553,
0.10460866242647171,
0.04732521250844002,
-0.11921124905347824,
0.009752476587891579,
0.09930381178855896,
-0.04945680499076843,
0.09620221704244614,
0.040187425911426544,
-0.14773541688919067,
0.036219459027051926,
-0.05416115000844002,
-0.06594996899366379,
-0.04819157347083092,
-0.044485192745923996,
-0.0397806353867054,
0.1325996071100235,
0.2043590396642685,
-0.046504415571689606,
0.04079987853765488,
-0.06612547487020493,
0.04139667749404907,
0.07320880144834518,
0.06640058755874634,
-0.04213016852736473,
-0.23575322329998016,
0.03131823241710663,
0.07714627683162689,
-0.023617465049028397,
-0.21513117849826813,
-0.08816151320934296,
0.009371439926326275,
-0.04920193552970886,
-0.06988794356584549,
0.10096220672130585,
0.11250386387109756,
0.05868158116936684,
-0.049654245376586914,
-0.1345376819372177,
-0.07777844369411469,
0.1739070564508438,
-0.14767830073833466,
-0.09548629075288773
] |
null | null | null |
# **Reinforce** Agent playing **Pixelcopter-PLE-v0**
This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
| {"tags": ["Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class"], "model-index": [{"name": "Reinforce-Pixelcopter-PLE-v0-3-layer-mlp-v0", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Pixelcopter-PLE-v0", "type": "Pixelcopter-PLE-v0"}, "metrics": [{"type": "mean_reward", "value": "24.70 +/- 27.06", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | TanHanlin/Reinforce-Pixelcopter-PLE-v0-3-layer-mlp-v0 | [
"Pixelcopter-PLE-v0",
"reinforce",
"reinforcement-learning",
"custom-implementation",
"deep-rl-class",
"model-index",
"region:us"
] | 2024-02-06T07:06:12+00:00 | [] | [] | TAGS
#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us
|
# Reinforce Agent playing Pixelcopter-PLE-v0
This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL
| [
"# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
"TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n",
"# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
41,
58
] | [
"passage: TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL"
] | [
0.0073175891302526,
-0.2259262204170227,
-0.0017347558168694377,
0.05054566636681557,
0.0658537745475769,
-0.055378563702106476,
0.1412602812051773,
0.05916554853320122,
-0.04990595206618309,
0.059261854737997055,
0.14166708290576935,
0.03996060788631439,
0.022112762555480003,
0.1513713151216507,
0.09764605015516281,
-0.2469022423028946,
0.07438477873802185,
0.01641594059765339,
0.008152224123477936,
0.09583204984664917,
0.060265738517045975,
-0.1405058205127716,
0.037032704800367355,
-0.01332044042646885,
-0.13650871813297272,
0.0010478810872882605,
-0.021802188828587532,
-0.03625129908323288,
0.15681709349155426,
0.006844013463705778,
0.09602472931146622,
-0.001560068572871387,
0.06475798785686493,
-0.12438877671957016,
0.05466329678893089,
0.06455880403518677,
-0.06293967366218567,
0.058029334992170334,
-0.057374246418476105,
0.11959903687238693,
0.04641333222389221,
-0.01578129455447197,
0.054811324924230576,
0.010941818356513977,
-0.14131468534469604,
-0.006710252724587917,
0.007013716734945774,
0.15098218619823456,
0.1339312642812729,
0.01409265398979187,
-0.0014771400019526482,
0.1363491266965866,
-0.16774429380893707,
0.045684073120355606,
0.061802688986063004,
-0.2633039951324463,
-0.04168876260519028,
0.12259352207183838,
0.08951573073863983,
0.06848238408565521,
-0.060910262167453766,
0.07636868953704834,
0.049813780933618546,
0.013985024765133858,
0.023094501346349716,
-0.042509064078330994,
-0.040479615330696106,
0.02289252169430256,
-0.0921095609664917,
-0.05999262258410454,
0.11517233401536942,
-0.006806366611272097,
0.03735918551683426,
-0.12476086616516113,
-0.015330453403294086,
-0.07314357161521912,
-0.05917041376233101,
-0.082573801279068,
0.07563583552837372,
0.030191516503691673,
-0.048283837735652924,
-0.08895846456289291,
-0.056533291935920715,
-0.11489585787057877,
-0.023082571104168892,
-0.07226225733757019,
0.005096882116049528,
-0.03157244250178337,
-0.035645097494125366,
0.09446526318788528,
-0.0021088174544274807,
-0.015028090216219425,
-0.03452150896191597,
-0.05930153280496597,
-0.04213470220565796,
-0.02359505370259285,
-0.03510070592164993,
-0.059062156826257706,
0.054655663669109344,
0.0680202916264534,
0.04938843473792076,
0.09133565425872803,
-0.0467856265604496,
0.1667373925447464,
-0.03256719931960106,
0.08078566938638687,
-0.011897698976099491,
0.2012830525636673,
0.11370102316141129,
0.12129533290863037,
0.06716908514499664,
-0.05294690653681755,
-0.16726544499397278,
0.039163749665021896,
0.12641896307468414,
0.07664673775434494,
-0.032492902129888535,
0.018162984400987625,
-0.12440363317728043,
0.05439428985118866,
-0.14826108515262604,
-0.06745084375143051,
0.024251462891697884,
0.01822635903954506,
-0.060682263225317,
0.03656952083110809,
-0.0028792342636734247,
0.003339326474815607,
0.004654870834201574,
-0.16432709991931915,
-0.05568019300699234,
0.028964387252926826,
-0.15712425112724304,
-0.06656725704669952,
0.06277995556592941,
-0.10113482922315598,
-0.012132617644965649,
-0.16982388496398926,
-0.16305199265480042,
-0.03628521412611008,
0.017857929691672325,
-0.040613796561956406,
-0.056917786598205566,
-0.14010562002658844,
-0.019415250048041344,
-0.045320261269807816,
-0.004312154371291399,
0.044072363525629044,
0.0020940210670232773,
0.04635847359895706,
0.0066573889926075935,
0.09289347380399704,
0.010714372619986534,
-0.0014722738415002823,
-0.04595406726002693,
0.0909833237528801,
-0.30731555819511414,
0.07525643706321716,
-0.08645553886890411,
0.05539081245660782,
-0.057316381484270096,
-0.0926317572593689,
-0.007509906310588121,
0.06277763843536377,
0.060464419424533844,
0.20788121223449707,
-0.2800109386444092,
-0.07025618106126785,
0.13655538856983185,
-0.09533236175775528,
-0.13146020472049713,
0.0513952374458313,
-0.050213608890771866,
0.07593657076358795,
0.027370907366275787,
0.140700101852417,
-0.028026295825839043,
-0.15554022789001465,
0.06281048059463501,
0.04586128890514374,
-0.11356306821107864,
0.019295670092105865,
0.03597676753997803,
0.06723599135875702,
0.05744141340255737,
-0.036986757069826126,
-0.04105675220489502,
0.08096802979707718,
-0.07076814025640488,
-0.037564266473054886,
0.04588831216096878,
-0.0579565204679966,
0.1630958467721939,
0.033971156924963,
0.09856503456830978,
-0.04149768501520157,
-0.07435470074415207,
-0.005698562134057283,
0.038746561855077744,
-0.08962973952293396,
0.025353478267788887,
-0.18320298194885254,
0.2423991560935974,
-0.02621818706393242,
0.027546977624297142,
-0.16845986247062683,
-0.0588528998196125,
0.011087946593761444,
0.21568740904331207,
0.030399197712540627,
0.12989304959774017,
0.07485637813806534,
-0.01250512059777975,
0.014156299643218517,
-0.06183977797627449,
-0.1972363442182541,
-0.03247830644249916,
0.008314179256558418,
-0.058311350643634796,
-0.04934588819742203,
-0.0900716632604599,
0.10427892208099365,
-0.19334633648395538,
-0.005319371819496155,
0.08282599598169327,
0.023504555225372314,
0.03946567326784134,
0.0035407328978180885,
-0.03634254261851311,
0.055148303508758545,
0.02030518464744091,
-0.08980578929185867,
0.14668866991996765,
0.0035520538222044706,
-0.03514726087450981,
-0.03927676007151604,
-0.03267495706677437,
0.05703731253743172,
0.08045367896556854,
-0.18214593827724457,
-0.0733821839094162,
-0.0838410034775734,
-0.02458474040031433,
0.050523869693279266,
0.036679428070783615,
0.02738112211227417,
0.44813573360443115,
0.057562243193387985,
0.09003535658121109,
-0.08811535686254501,
0.039806611835956573,
0.012785476632416248,
-0.031281858682632446,
0.013625281862914562,
0.04725322127342224,
0.11279468983411789,
0.028284218162298203,
0.01669839769601822,
0.03680038824677467,
0.01938779093325138,
0.08824212104082108,
-0.10939645022153854,
-0.003965397831052542,
0.002614045049995184,
0.038018375635147095,
0.03672022372484207,
0.07190682739019394,
0.015936892479658127,
-0.09583546966314316,
-0.030848123133182526,
-0.11166880279779434,
0.015594755299389362,
-0.20979784429073334,
-0.025905707851052284,
-0.029619399458169937,
0.0003502996696624905,
0.09109684824943542,
0.04222718998789787,
-0.04444896802306175,
0.035467714071273804,
0.03947039321064949,
-0.0861397460103035,
0.0594942644238472,
-0.014317752793431282,
-0.07008631527423859,
0.13023322820663452,
-0.1002996563911438,
-0.3153233230113983,
-0.08797995746135712,
0.05698639526963234,
0.05295826122164726,
0.06816939264535904,
-0.05876303091645241,
-0.09240786731243134,
0.03294730558991432,
-0.06836386770009995,
-0.0017794050509110093,
0.0037346978206187487,
-0.051060982048511505,
0.07253886014223099,
0.08541567623615265,
-0.014505518600344658,
-0.08911184966564178,
-0.006620637606829405,
-0.041561197489500046,
-0.124965138733387,
0.044060997664928436,
-0.03760828450322151,
0.00007921225915197283,
0.18620672821998596,
0.03724536672234535,
0.06256633251905441,
-0.06291008740663528,
0.07596296072006226,
-0.09150096774101257,
0.0004740063741337508,
0.18428465723991394,
-0.015377625823020935,
-0.004100616089999676,
-0.03996327146887779,
-0.0259257685393095,
-0.10829219967126846,
0.053985193371772766,
-0.07330703735351562,
-0.07349077612161636,
-0.0023273853585124016,
-0.07770214974880219,
-0.0351552739739418,
0.0012160884216427803,
0.07817990332841873,
0.029699061065912247,
-0.09635239094495773,
0.04920589178800583,
0.1298678070306778,
0.0931883230805397,
0.03626195341348648,
0.023981640115380287,
0.13739009201526642,
-0.11230582743883133,
0.019063033163547516,
-0.05148853361606598,
-0.1041760966181755,
-0.042787205427885056,
-0.0714287981390953,
0.07368279993534088,
0.06034531816840172,
-0.09970010071992874,
0.05144011229276657,
0.041872985661029816,
0.0883496031165123,
0.1373600959777832,
-0.04213863983750343,
-0.11244629323482513,
-0.041393622756004333,
-0.022004956379532814,
-0.1777329444885254,
0.0341336652636528,
0.22155584394931793,
0.0073304991237819195,
-0.10497386753559113,
0.07876885682344437,
-0.005956185050308704,
0.11527370661497116,
0.031222699210047722,
-0.278682678937912,
0.016931315883994102,
0.00203216471709311,
0.042359162122011185,
-0.047676295042037964,
0.10937416553497314,
0.11747439950704575,
-0.14421136677265167,
-0.06650938838720322,
-0.03273930773139,
0.044137366116046906,
-0.15618287026882172,
0.036923591047525406,
-0.12602220475673676,
0.06240779533982277,
0.050940994173288345,
0.05090156942605972,
-0.2197665423154831,
0.06881614029407501,
-0.0274215005338192,
0.06763827055692673,
-0.062248338013887405,
-0.01823522336781025,
0.04473711550235748,
0.025079863145947456,
0.14955177903175354,
-0.014347962103784084,
0.14454017579555511,
-0.09031219780445099,
-0.11753576993942261,
0.0027052261866629124,
0.08532248437404633,
0.013173088431358337,
0.013580933213233948,
0.0026939227245748043,
0.041669201105833054,
-0.02811569906771183,
0.17063532769680023,
-0.08147624880075455,
-0.022407781332731247,
-0.06592555344104767,
-0.018158966675400734,
0.2039334923028946,
-0.12064731866121292,
-0.10121093690395355,
-0.11619500070810318,
0.08663272857666016,
-0.04296411573886871,
0.08175522089004517,
-0.020344657823443413,
0.049704354256391525,
-0.02509051002562046,
0.007178863976150751,
0.09594997018575668,
0.01950966566801071,
0.08983828872442245,
-0.09791163355112076,
-0.019585272297263145,
0.13838915526866913,
-0.037155888974666595,
-0.036971647292375565,
-0.019425252452492714,
0.11054370552301407,
-0.0358734093606472,
0.08033111691474915,
0.03929615020751953,
0.03664831817150116,
0.03428546339273453,
-0.039165496826171875,
0.10309428721666336,
0.10041618347167969,
-0.06291446089744568,
0.03864621743559837,
-0.07954532653093338,
0.26597461104393005,
0.040773067623376846,
0.07301845401525497,
0.28390514850616455,
0.19391325116157532,
-0.03036464750766754,
0.10683353990316391,
-0.017607249319553375,
-0.024403288960456848,
-0.2950931787490845,
0.0006976581644266844,
0.027765681967139244,
0.11812873929738998,
0.01744898222386837,
-0.20587195456027985,
-0.1211688369512558,
-0.03560304269194603,
-0.007791717536747456,
0.0310499370098114,
-0.2441052496433258,
-0.06442268192768097,
0.06107868626713753,
0.13779635727405548,
0.15878525376319885,
-0.05917542055249214,
-0.007856467738747597,
0.029358724132180214,
0.07593556493520737,
0.017292039468884468,
-0.11598441749811172,
0.11550791561603546,
0.025637371465563774,
-0.05708931386470795,
0.0267958827316761,
-0.044003549963235855,
0.04214555397629738,
-0.17736166715621948,
0.10933554917573929,
-0.05924695357680321,
-0.08421005308628082,
0.07140472531318665,
-0.02217724733054638,
-0.048552993685007095,
0.0789642184972763,
0.020652711391448975,
-0.13173207640647888,
0.038154006004333496,
0.005618774797767401,
0.04346654564142227,
-0.004941361024975777,
-0.019811764359474182,
-0.029163256287574768,
0.07706235349178314,
-0.03806605935096741,
0.09605937451124191,
0.19590972363948822,
-0.0573095865547657,
0.03974950686097145,
0.085201695561409,
0.09593135863542557,
-0.05523005872964859,
-0.0809539332985878,
-0.03812742978334427,
-0.005277194548398256,
0.0674438327550888,
-0.08598461747169495,
-0.019085103645920753,
0.07938229292631149,
0.015313901007175446,
0.14910826086997986,
0.14389736950397491,
-0.08835655450820923,
0.11321785300970078,
0.10694554448127747,
-0.11366690695285797,
-0.08583837002515793,
-0.02963297814130783,
0.0009990704711526632,
0.04910186678171158,
-0.048617590218782425,
0.05932905897498131,
-0.1035301461815834,
0.012819357216358185,
0.03532040864229202,
0.0038119733799248934,
-0.09975302964448929,
0.009764863178133965,
0.08645275235176086,
0.06119582802057266,
-0.0567571222782135,
0.09250631928443909,
-0.0019178141374140978,
-0.10868195444345474,
0.07241881638765335,
0.009918469935655594,
-0.021528873592615128,
-0.06352251768112183,
0.03211374953389168,
0.2370220273733139,
0.13945111632347107,
-0.04336636886000633,
-0.12396618723869324,
-0.15508891642093658,
0.037849195301532745,
0.024356422945857048,
0.051251959055662155,
0.0062240250408649445,
-0.06906022876501083,
0.01234503649175167,
-0.04392383247613907,
0.005266309250146151,
-0.05930564925074577,
-0.047703344374895096,
-0.12081446498632431,
0.1154373437166214,
0.053290288895368576,
0.11705748736858368,
-0.0842847004532814,
-0.07057584822177887,
-0.1921386867761612,
0.09190598875284195,
0.041707299649715424,
-0.05532265454530716,
0.06002674251794815,
-0.030134430155158043,
0.017344338819384575,
0.11256659775972366,
-0.051967836916446686,
0.008543911390006542,
-0.09269233793020248,
0.03236149623990059,
0.03133073076605797,
0.04903566092252731,
-0.004612727556377649,
-0.017903391271829605,
0.04399999976158142,
-0.05730267986655235,
0.07619527727365494,
-0.07757602632045746,
-0.033709146082401276,
0.0645759105682373,
-0.16051416099071503,
-0.054324716329574585,
0.08708633482456207,
0.013749903067946434,
0.02590017393231392,
-0.05825240537524223,
0.019142305478453636,
-0.05566488951444626,
-0.04483235627412796,
0.01169554702937603,
-0.05552767962217331,
-0.011517677456140518,
0.05293213203549385,
-0.05287189036607742,
-0.040493328124284744,
-0.06794002652168274,
0.061874233186244965,
-0.07247710227966309,
0.09816460311412811,
0.031187955290079117,
-0.10892423242330551,
0.07648903876543045,
-0.037552736699581146,
-0.0049397205002605915,
-0.009439278393983841,
0.039307788014411926,
0.15598824620246887,
-0.1606634259223938,
0.05345672369003296,
-0.0484454482793808,
0.13272921741008759,
0.046888746321201324,
-0.04458791762590408,
-0.020207170397043228,
0.02469455823302269,
-0.05549024045467377,
0.06932897865772247,
0.15877580642700195,
0.09880131483078003,
0.02571805939078331,
0.008134597912430763,
0.10187267512083054,
0.1060529574751854,
0.08136752992868423,
0.08394161611795425,
-0.03428563475608826,
-0.11287897825241089,
0.14338994026184082,
0.09748584777116776,
0.024613093584775925,
0.21077860891819,
0.17944025993347168,
0.03125298395752907,
0.03018142655491829,
-0.06512103229761124,
0.17325744032859802,
0.061261482536792755,
-0.08229418843984604,
0.014424329623579979,
0.03221147879958153,
-0.049809664487838745,
-0.047004032880067825,
-0.09757380187511444,
-0.029556652531027794,
-0.24085633456707,
0.10851483792066574,
-0.057250600308179855,
-0.09750643372535706,
0.022772664204239845,
0.02990041859447956,
-0.018839845433831215,
0.11280566453933716,
-0.07735858112573624,
0.012980576604604721,
0.18577688932418823,
-0.03825045004487038,
-0.022322099655866623,
-0.1633504331111908,
-0.11154003441333771,
-0.014046176336705685,
-0.11750495433807373,
0.025494296103715897,
0.06305963546037674,
0.01117965579032898,
0.04399528726935387,
0.028923438861966133,
-0.020834028720855713,
0.019218796864151955,
-0.05903913825750351,
-0.042673509567976,
-0.01891910657286644,
0.02202831581234932,
-0.09593231230974197,
-0.03627033904194832,
0.12151803076267242,
-0.03246605768799782,
-0.08207374066114426,
-0.006544890813529491,
0.07848484069108963,
-0.042620159685611725,
0.09450104832649231,
-0.07687012106180191,
-0.03479038178920746,
-0.06794454902410507,
0.268902063369751,
0.09388194978237152,
-0.20183001458644867,
0.03341769427061081,
-0.030470456928014755,
0.026735708117485046,
-0.09215684235095978,
0.16250114142894745,
0.0899243950843811,
0.049168527126312256,
-0.12686687707901,
-0.003401300171390176,
-0.09992645680904388,
-0.0028723697178065777,
-0.12552696466445923,
-0.14725084602832794,
0.12093491852283478,
-0.003848524997010827,
-0.06547791510820389,
0.02844911813735962,
-0.15909899771213531,
0.06585367769002914,
0.0978507474064827,
-0.1514272391796112,
-0.038227714598178864,
-0.06086801365017891,
0.06072385236620903,
0.026465637609362602,
0.13005392253398895,
-0.05080926790833473,
0.012067130766808987,
-0.0656723901629448,
-0.011309894733130932,
-0.0000654291216051206,
-0.017478201538324356,
0.001532604917883873,
-0.09828947484493256,
0.05038110539317131,
-0.0835796371102333,
0.12184429168701172,
0.05709611251950264,
0.005326167680323124,
0.008464806713163853,
0.0648408755660057,
-0.02414623089134693,
-0.10202058404684067,
-0.01877439208328724,
0.033475372940301895,
0.03998998552560806,
0.010373802855610847,
0.034506846219301224,
0.0006507808575406671,
0.07714920490980148,
-0.011413984932005405,
-0.027285432443022728,
-0.058209117501974106,
0.03936338797211647,
-0.10441672056913376,
0.10461361706256866,
0.0013552121818065643,
-0.02240127883851528,
-0.010913821868598461,
-0.05532446503639221,
0.045815300196409225,
0.04572062939405441,
0.029743505641818047,
-0.05261747166514397,
-0.09262793511152267,
-0.021781492978334427,
0.023900283500552177,
-0.11539579927921295,
-0.18497975170612335,
-0.0664035826921463,
-0.15038692951202393,
-0.01633414439857006,
-0.0620744526386261,
0.08902198076248169,
0.13558129966259003,
0.030392181128263474,
-0.04822919890284538,
-0.12171997129917145,
0.025026977062225342,
0.13544774055480957,
-0.03851630911231041,
-0.07532322406768799
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | shidowake/test-240206-cyber2base-7B-qlora-adaptor | [
"transformers",
"tensorboard",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-06T07:07:08+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #tensorboard #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #tensorboard #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
35,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05129265785217285,
0.22092880308628082,
-0.0035179967526346445,
0.02501223422586918,
0.11708144098520279,
-0.0005181307788006961,
0.04833335056900978,
0.126449316740036,
-0.019315702840685844,
0.11368890851736069,
0.02830384485423565,
0.087161585688591,
0.10932693630456924,
0.14976787567138672,
0.027767986059188843,
-0.22400330007076263,
0.013545092195272446,
-0.08547151833772659,
0.007360507268458605,
0.10887812823057175,
0.13407185673713684,
-0.10422329604625702,
0.0799589455127716,
-0.022218724712729454,
-0.01576193980872631,
-0.0024709447752684355,
-0.09191098809242249,
-0.07704412192106247,
0.06658186763525009,
0.07635242491960526,
0.06079627200961113,
0.015532677061855793,
0.09731730818748474,
-0.28775396943092346,
0.015315494500100613,
0.08426344394683838,
-0.003832674352452159,
0.06570416688919067,
0.06679647415876389,
-0.07225077599287033,
0.11294640600681305,
-0.08326761424541473,
0.14265748858451843,
0.07819409668445587,
-0.0751931369304657,
-0.20058366656303406,
-0.0678529441356659,
0.08777504414319992,
0.12172447144985199,
0.06566067785024643,
-0.0240190327167511,
0.15403629839420319,
-0.07573149353265762,
0.009909119457006454,
0.1344568133354187,
-0.09381069988012314,
-0.05284041166305542,
0.04851995036005974,
0.11218059062957764,
0.09264162927865982,
-0.13201914727687836,
0.007263765670359135,
0.03538952395319939,
0.01823991909623146,
0.08935598284006119,
0.021646643057465553,
0.10804788768291473,
0.0471319817006588,
-0.13886919617652893,
-0.04078420251607895,
0.10519677400588989,
0.03467073664069176,
-0.0511174350976944,
-0.21634697914123535,
-0.0042060064151883125,
-0.018993662670254707,
-0.02438262850046158,
-0.057221945375204086,
0.047075625509023666,
-0.03517723083496094,
0.05707012489438057,
-0.03823539987206459,
-0.09978390485048294,
-0.030010027810931206,
0.07668479532003403,
0.05997013673186302,
0.013257342390716076,
-0.018452148884534836,
0.029318852350115776,
0.11608750373125076,
0.04971613734960556,
-0.1120513305068016,
-0.06576516479253769,
-0.06617631763219833,
-0.1015271320939064,
-0.04605502262711525,
0.04043229669332504,
0.021388711407780647,
0.02790878526866436,
0.2005361020565033,
0.0013144337572157383,
0.04578043892979622,
0.027719371020793915,
0.006634909193962812,
0.06140429154038429,
0.09704186022281647,
-0.06184549257159233,
-0.14441433548927307,
-0.052150655537843704,
0.0838291123509407,
-0.005048004910349846,
-0.03949505463242531,
-0.05814170464873314,
0.045859213918447495,
0.050563037395477295,
0.11559558659791946,
0.0925697535276413,
-0.00952589139342308,
-0.04793645069003105,
-0.03162948787212372,
0.22493304312229156,
-0.14139817655086517,
0.04953770712018013,
-0.012962840497493744,
-0.03824609890580177,
-0.04221971333026886,
0.031711336225271225,
0.03209373727440834,
-0.016264338046312332,
0.09782443940639496,
-0.05990086868405342,
-0.0385088287293911,
-0.10335034132003784,
-0.04756024107336998,
0.03825374320149422,
-0.010136007331311703,
-0.014318512752652168,
-0.06374996900558472,
-0.0976734608411789,
-0.0440363809466362,
0.06589976698160172,
-0.06674116849899292,
-0.04861561208963394,
0.014166015200316906,
-0.05206536501646042,
-0.0034754190128296614,
-0.0007777506252750754,
0.11236779391765594,
-0.02925705723464489,
0.03520512953400612,
-0.03593742102384567,
0.06071564927697182,
0.10378561168909073,
0.04010351747274399,
-0.06594906002283096,
0.052586451172828674,
-0.21832634508609772,
0.09042318910360336,
-0.10633471608161926,
0.030845342203974724,
-0.15926258265972137,
-0.041567448526620865,
0.019635990262031555,
0.014180500991642475,
0.008806075900793076,
0.11377737671136856,
-0.19005726277828217,
-0.025171801447868347,
0.1300044059753418,
-0.0945831909775734,
-0.10084225982427597,
0.07147544622421265,
-0.04920754209160805,
0.13694678246974945,
0.04123294726014137,
-0.02705676108598709,
0.06769294291734695,
-0.15995272994041443,
-0.05838307738304138,
-0.016111139208078384,
-0.009447884745895863,
0.13041087985038757,
0.05866660177707672,
-0.059992723166942596,
0.08056262880563736,
0.02419431507587433,
-0.019690977409482002,
-0.043121110647916794,
-0.043974619358778,
-0.10198459774255753,
-0.0013167005963623524,
-0.07957979291677475,
0.04423648864030838,
-0.008130980655550957,
-0.07807544618844986,
-0.029026441276073456,
-0.1834263950586319,
0.052552398294210434,
0.08257346600294113,
0.016371481120586395,
-0.010266339406371117,
-0.07969299703836441,
0.008663319051265717,
-0.026843959465622902,
-0.016039082780480385,
-0.1699541211128235,
-0.04567434638738632,
0.04240812361240387,
-0.1652011275291443,
0.04373474419116974,
-0.03978709131479263,
0.05944959074258804,
0.0320458859205246,
-0.05190446972846985,
-0.0011512517230585217,
-0.014688361436128616,
0.01982065849006176,
-0.036412887275218964,
-0.19573865830898285,
-0.043008726090192795,
-0.02956235036253929,
0.160308837890625,
-0.24183565378189087,
0.03329772129654884,
0.057998064905405045,
0.1498682051897049,
0.0007596806390210986,
-0.04587903618812561,
0.02215476520359516,
-0.05358593538403511,
-0.05101087689399719,
-0.06615893542766571,
-0.0030418226961046457,
-0.028946032747626305,
-0.04868051037192345,
0.029185006394982338,
-0.1843569129705429,
-0.04010150954127312,
0.10059087723493576,
0.09962263703346252,
-0.1534094512462616,
-0.014867037534713745,
-0.04807868227362633,
-0.06743370741605759,
-0.08563356846570969,
-0.05508554354310036,
0.14246419072151184,
0.05223851278424263,
0.04553950950503349,
-0.07807126641273499,
-0.0671616643667221,
0.01849428191781044,
0.0007367139332927763,
-0.040424644947052,
0.07459479570388794,
0.08395206183195114,
-0.09447163343429565,
0.0763159990310669,
0.07563351839780807,
0.0745408833026886,
0.10207726806402206,
0.01656539924442768,
-0.10977036505937576,
-0.025348074734210968,
0.015244545415043831,
0.024517716839909554,
0.14840243756771088,
-0.05424882099032402,
0.03563909977674484,
0.050164610147476196,
-0.04687408357858658,
0.02342480979859829,
-0.10203663259744644,
0.031017832458019257,
0.039212487637996674,
-0.0007507543195970356,
0.04622079059481621,
-0.04102898761630058,
0.00785372406244278,
0.0762152150273323,
0.049299295991659164,
0.04492497444152832,
0.00451761344447732,
-0.01614830642938614,
-0.09929052740335464,
0.16183419525623322,
-0.09384983777999878,
-0.3009108603000641,
-0.15124821662902832,
0.016429319977760315,
0.03624126687645912,
-0.02517835609614849,
0.026567041873931885,
-0.05863599851727486,
-0.10570778697729111,
-0.10092674195766449,
-0.005781888496130705,
0.020709572359919548,
-0.07780015468597412,
-0.06983298063278198,
0.07268352061510086,
0.03807120397686958,
-0.14607851207256317,
0.03948947787284851,
0.04788151755928993,
-0.050529371947050095,
-0.02089620567858219,
0.08623623847961426,
0.11477750539779663,
0.15922853350639343,
-0.016451021656394005,
-0.02623858116567135,
0.018677379935979843,
0.2000662386417389,
-0.13754040002822876,
0.11520840972661972,
0.13333985209465027,
-0.047717101871967316,
0.09115317463874817,
0.17400066554546356,
0.026365477591753006,
-0.07634809613227844,
0.037423089146614075,
0.050920601934194565,
-0.04808442294597626,
-0.25332513451576233,
-0.05663376301527023,
0.009860052727162838,
-0.08193530142307281,
0.093865767121315,
0.09549196064472198,
0.14020170271396637,
0.03696327283978462,
-0.07213930040597916,
-0.042459435760974884,
0.0036775453481823206,
0.11384411156177521,
-0.03568359836935997,
-0.005343656521290541,
0.08481227606534958,
-0.04323641583323479,
-0.0007266122847795486,
0.10209894180297852,
0.025129051879048347,
0.18802158534526825,
0.023904064670205116,
0.1413942575454712,
0.06004384160041809,
0.06251177191734314,
0.000551602803170681,
0.014730905182659626,
0.0419386588037014,
0.014600582420825958,
-0.007120285648852587,
-0.09679079800844193,
0.014006711542606354,
0.13750481605529785,
0.05959208309650421,
0.020287446677684784,
0.011764034628868103,
-0.022532951086759567,
0.05339716002345085,
0.17398898303508759,
-0.014597101137042046,
-0.19965122640132904,
-0.06684007495641708,
0.07284105569124222,
-0.05776670575141907,
-0.12176336348056793,
-0.03483381122350693,
0.030525391921401024,
-0.17502301931381226,
0.03214195370674133,
-0.02044833078980446,
0.10361938923597336,
-0.10804998129606247,
-0.029580531641840935,
0.02396276406943798,
0.08289678394794464,
-0.021593092009425163,
0.09409880638122559,
-0.15318867564201355,
0.12421330064535141,
0.027643848210573196,
0.08287007361650467,
-0.11865443736314774,
0.0884069874882698,
-0.00914052128791809,
0.004794934764504433,
0.17335905134677887,
-0.00887978170067072,
-0.06141883134841919,
-0.06713027507066727,
-0.0922374352812767,
-0.02362675592303276,
0.11753249168395996,
-0.10957267880439758,
0.08096589893102646,
-0.00771739799529314,
-0.049180056899785995,
0.013477327302098274,
-0.11567004770040512,
-0.16323277354240417,
-0.20408131182193756,
0.07107069343328476,
-0.0975819006562233,
-0.0031133198644965887,
-0.10442767292261124,
-0.06875836104154587,
-0.031984321773052216,
0.23626349866390228,
-0.13814060389995575,
-0.079515740275383,
-0.15315008163452148,
-0.05907658115029335,
0.17421935498714447,
-0.038646843284368515,
0.07752027362585068,
-0.009368468075990677,
0.22135040163993835,
-0.0005769561394117773,
-0.006669431459158659,
0.06871405988931656,
-0.08641299605369568,
-0.16795587539672852,
-0.0766071230173111,
0.13526922464370728,
0.11764655262231827,
0.052490342408418655,
-0.005320967175066471,
0.010855535045266151,
-0.025631245225667953,
-0.1058010458946228,
0.001448161550797522,
0.1267717480659485,
0.057947464287281036,
0.022029021754860878,
-0.03157263621687889,
-0.10811471194028854,
-0.06623940169811249,
-0.052098218351602554,
0.04786430671811104,
0.181030735373497,
-0.09767430275678635,
0.17944495379924774,
0.14716151356697083,
-0.06651145964860916,
-0.20653264224529266,
0.04529459774494171,
0.048565998673439026,
-0.011420371010899544,
0.04028045013546944,
-0.18920361995697021,
0.08407842367887497,
0.012931931763887405,
-0.054178349673748016,
0.13587801158428192,
-0.1596759408712387,
-0.1581452637910843,
0.06423576176166534,
0.045448124408721924,
-0.23230639100074768,
-0.1363905370235443,
-0.08674333989620209,
-0.062469057738780975,
-0.1569693386554718,
0.08063073456287384,
-0.00359702599234879,
0.007655099034309387,
0.04580266401171684,
0.029607584699988365,
0.021776875481009483,
-0.05363304913043976,
0.18664778769016266,
-0.0077520408667624,
0.013649445958435535,
-0.07357750833034515,
-0.07677493244409561,
0.09593275189399719,
-0.05751553177833557,
0.10879078507423401,
-0.0018659495981410146,
0.008710069581866264,
-0.09187035262584686,
-0.056738369166851044,
-0.04518502950668335,
0.05294037610292435,
-0.08283393830060959,
-0.11107265949249268,
-0.04874441400170326,
0.08882616460323334,
0.08018611371517181,
-0.03557325154542923,
-0.012883543968200684,
-0.07679562270641327,
0.08622492104768753,
0.19816941022872925,
0.16589786112308502,
0.02310795709490776,
-0.08369841426610947,
0.011619039811193943,
-0.03282969444990158,
0.03419959545135498,
-0.23653645813465118,
0.037609003484249115,
0.05255814641714096,
0.035989757627248764,
0.10626954585313797,
-0.025342373177409172,
-0.1775771826505661,
-0.04355216026306152,
0.060009121894836426,
-0.043581295758485794,
-0.21388691663742065,
-0.013861402869224548,
0.0937437117099762,
-0.19061240553855896,
-0.02898477576673031,
0.026441190391778946,
-0.0349886454641819,
-0.03127404302358627,
0.005819133948534727,
0.060078177601099014,
0.027275938540697098,
0.08815191686153412,
0.06973249465227127,
0.09625289589166641,
-0.0921335518360138,
0.1012224555015564,
0.10827095806598663,
-0.09258020669221878,
0.03437352553009987,
0.06997713446617126,
-0.0464654341340065,
-0.03779982030391693,
0.04445716738700867,
0.05210983380675316,
-0.0035824330989271402,
-0.05700870603322983,
-0.0012550450628623366,
-0.05801399052143097,
0.053437739610672,
0.11578024923801422,
0.024411514401435852,
-0.031481921672821045,
0.06297732889652252,
0.031466688960790634,
-0.11282244324684143,
0.09800126403570175,
0.016209103167057037,
0.0332130491733551,
-0.057540472596883774,
-0.014383697882294655,
0.04758455976843834,
0.01908467337489128,
-0.018431948497891426,
-0.03139050304889679,
-0.03864758461713791,
-0.01871557906270027,
-0.15688249468803406,
-0.00880077388137579,
-0.06904306262731552,
0.008402650244534016,
0.00812726654112339,
-0.043084751814603806,
-0.007121704053133726,
0.027271781116724014,
-0.07522907853126526,
-0.06459254771471024,
-0.004998138640075922,
0.09239296615123749,
-0.15713782608509064,
0.0025221933610737324,
0.0771171823143959,
-0.1061064675450325,
0.06508705765008926,
-0.008114354684948921,
0.0024605898652225733,
0.013164390809834003,
-0.14945632219314575,
0.05327258259057999,
-0.013698753900825977,
0.019031856209039688,
0.04311361908912659,
-0.16456228494644165,
0.004066106863319874,
-0.04824318736791611,
-0.02670198306441307,
-0.010943911038339138,
-0.06265503913164139,
-0.11827512085437775,
0.0809461921453476,
-0.014609033241868019,
-0.06019756942987442,
-0.01396041177213192,
0.05710810795426369,
0.09468782693147659,
-0.03689182177186012,
0.09434254467487335,
-0.001471250900067389,
0.06198366731405258,
-0.17185549437999725,
-0.025822067633271217,
-0.035810794681310654,
0.014255421236157417,
0.023194478824734688,
-0.010762988589704037,
0.04005304351449013,
-0.005854793358594179,
0.23268097639083862,
-0.031446292996406555,
0.15102393925189972,
0.05785702168941498,
0.00336667918600142,
0.0021511006634682417,
0.06947512179613113,
0.05804704874753952,
0.026832543313503265,
0.007880797609686852,
0.029525361955165863,
-0.023783985525369644,
-0.011394080705940723,
-0.16318662464618683,
0.03503057733178139,
0.13904942572116852,
0.0770760327577591,
0.010082700289785862,
0.07332480698823929,
-0.12486710399389267,
-0.11119920760393143,
0.10134211927652359,
-0.03071645461022854,
0.009480697102844715,
-0.07685033231973648,
0.13950638473033905,
0.14809831976890564,
-0.15144896507263184,
0.069137804210186,
-0.05071515217423439,
-0.05081162974238396,
-0.09150747954845428,
-0.11238798499107361,
-0.06143717095255852,
-0.03477819636464119,
0.002448677783831954,
-0.04411304369568825,
0.05612285062670708,
0.050102539360523224,
-0.009517417289316654,
0.0061646318063139915,
0.11181281507015228,
-0.010644787922501564,
-0.0011640603188425303,
0.0354636125266552,
0.040780141949653625,
0.02505035698413849,
-0.0645608976483345,
0.03148636221885681,
0.017563825473189354,
0.036894720047712326,
0.059959281235933304,
0.02824280969798565,
-0.036668822169303894,
0.028542157262563705,
0.007808319292962551,
-0.10596310347318649,
0.023117657750844955,
-0.01790459267795086,
-0.06492031365633011,
0.12684021890163422,
0.03216429799795151,
0.014280314557254314,
-0.04237029701471329,
0.23208919167518616,
-0.06517360359430313,
-0.07358380407094955,
-0.13262179493904114,
0.09425038844347,
-0.020774539560079575,
0.05422547087073326,
0.03925251588225365,
-0.1250264197587967,
0.0033500357531011105,
0.13577201962471008,
0.11836834996938705,
-0.002868342911824584,
0.008071486838161945,
0.039927463978528976,
0.005613021552562714,
-0.05908596143126488,
0.036810338497161865,
0.05790701508522034,
0.13603076338768005,
-0.07816687971353531,
0.06294257938861847,
0.0068999994546175,
-0.07861924916505814,
-0.0430021770298481,
0.12258011102676392,
-0.01771291345357895,
0.033160895109176636,
-0.04474131762981415,
0.10494659841060638,
-0.06386210769414902,
-0.3023519814014435,
0.03618663549423218,
-0.10097576677799225,
-0.1573091298341751,
-0.01591421850025654,
0.05504710227251053,
-0.024499690160155296,
0.024731650948524475,
0.07404778897762299,
-0.05839020013809204,
0.18247833847999573,
0.03708400949835777,
-0.08908724039793015,
-0.05214424058794975,
0.05819600820541382,
-0.07667914777994156,
0.29527547955513,
0.0018856418319046497,
0.029907740652561188,
0.10656348615884781,
-0.02130613848567009,
-0.15849106013774872,
0.017152920365333557,
0.10987870395183563,
-0.09249646216630936,
0.08097189664840698,
0.1969153881072998,
-0.017194712534546852,
0.11304406821727753,
0.061918240040540695,
-0.060775525867938995,
0.0563589483499527,
-0.06079897657036781,
-0.04729166626930237,
-0.09214385598897934,
0.06507396697998047,
-0.05960074067115784,
0.15478351712226868,
0.0983923152089119,
-0.04626932740211487,
-0.011300264857709408,
-0.05205517262220383,
0.037209995090961456,
0.013506266288459301,
0.12517105042934418,
0.005834320094436407,
-0.16502192616462708,
0.03574982285499573,
-0.008094683289527893,
0.1093440055847168,
-0.24298563599586487,
-0.0835915356874466,
0.08754356950521469,
-0.028303591534495354,
-0.04819139838218689,
0.0978732705116272,
0.07554362714290619,
0.044536542147397995,
-0.046234432607889175,
-0.08701527863740921,
-0.015517168678343296,
0.14954860508441925,
-0.14008408784866333,
-0.008103124797344208
] |
null | null | transformers |
# Model Card for Model ID
merge(ties) model
* base model for ENERGY-DRINK-LOVE/SOLAR_merge2_dpo | {"license": "apache-2.0", "library_name": "transformers"} | text-generation | ENERGY-DRINK-LOVE/SOLAR_merge2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T07:07:28+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
merge(ties) model
* base model for ENERGY-DRINK-LOVE/SOLAR_merge2_dpo | [
"# Model Card for Model ID\n\nmerge(ties) model\n\n* base model for ENERGY-DRINK-LOVE/SOLAR_merge2_dpo"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID\n\nmerge(ties) model\n\n* base model for ENERGY-DRINK-LOVE/SOLAR_merge2_dpo"
] | [
59,
34
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID\n\nmerge(ties) model\n\n* base model for ENERGY-DRINK-LOVE/SOLAR_merge2_dpo"
] | [
-0.05547250062227249,
-0.003707007272168994,
-0.0045189084485173225,
-0.007060525473207235,
0.08705079555511475,
0.021368956193327904,
0.20419573783874512,
0.029576925560832024,
0.06726778298616409,
-0.024634862318634987,
0.10710130631923676,
0.17125500738620758,
0.04742515832185745,
0.1649344563484192,
-0.06739622354507446,
-0.15354694426059723,
0.14162424206733704,
0.043411701917648315,
0.008861851878464222,
0.0635051354765892,
0.13338401913642883,
-0.02757152169942856,
0.09661448001861572,
-0.036647979170084,
-0.10551515221595764,
0.029409293085336685,
-0.03491221368312836,
-0.06513315439224243,
0.10412147641181946,
0.050304777920246124,
0.0878150686621666,
0.10362210869789124,
-0.011331036686897278,
-0.12379325926303864,
0.03432922065258026,
-0.008170470595359802,
-0.07869963347911835,
0.07580269128084183,
0.032668933272361755,
-0.05190735682845116,
0.12746471166610718,
0.013509014621376991,
0.03820943459868431,
0.023575928062200546,
-0.031135203316807747,
-0.1342754364013672,
-0.02832230180501938,
0.0871984139084816,
0.10379648953676224,
0.017092756927013397,
0.04183001443743706,
0.08789791911840439,
0.04015873745083809,
0.044381722807884216,
-0.031133605167269707,
-0.1338079422712326,
0.06176162511110306,
0.18959812819957733,
-0.014348536729812622,
-0.024350935593247414,
0.07351797819137573,
0.0528084971010685,
0.06526073068380356,
-0.018719302490353584,
0.09453940391540527,
-0.00022627889120485634,
0.030029918998479843,
0.05154348537325859,
-0.08413555473089218,
-0.039855483919382095,
0.1997198909521103,
0.034958887845277786,
-0.035404838621616364,
-0.10586286336183548,
-0.1482587456703186,
0.044119805097579956,
-0.0481446273624897,
-0.02186594530940056,
0.0361744649708271,
0.07615634799003601,
0.10150770843029022,
0.009968679398298264,
-0.0856606662273407,
-0.02836696431040764,
-0.168846994638443,
0.22672954201698303,
0.0425662063062191,
0.06253346800804138,
-0.07503611594438553,
0.11924150586128235,
0.028788376599550247,
-0.10604243725538254,
-0.014095637947320938,
-0.08168215304613113,
0.0703907161951065,
0.016557589173316956,
-0.09346306324005127,
-0.024322550743818283,
0.09397207945585251,
0.1864035278558731,
0.07648760825395584,
-0.04819508641958237,
0.06744774430990219,
0.045549504458904266,
0.07084241509437561,
-0.07913445681333542,
-0.012017832137644291,
-0.051636070013046265,
0.03596346080303192,
-0.06539499014616013,
0.13027946650981903,
-0.02577095478773117,
-0.14240747690200806,
0.04182257130742073,
-0.05978085845708847,
0.0360632985830307,
0.023466255515813828,
0.06786546111106873,
-0.05474555119872093,
-0.01733436807990074,
0.21881593763828278,
-0.041188474744558334,
-0.04154471307992935,
0.005960405338555574,
0.04892769455909729,
0.15143941342830658,
0.11395581066608429,
0.04702756926417351,
0.05339282751083374,
-0.002391170710325241,
-0.025101253762841225,
-0.08194155246019363,
-0.09442263841629028,
-0.045596327632665634,
0.02925674244761467,
-0.004417967051267624,
0.11777845025062561,
-0.18711918592453003,
-0.30906423926353455,
0.08111893385648727,
0.06473273038864136,
-0.06605973839759827,
-0.0966896265745163,
-0.03091225028038025,
0.036863572895526886,
-0.018330128863453865,
-0.017332155257463455,
0.02439839392900467,
-0.019908223301172256,
-0.0035157320089638233,
0.053829699754714966,
0.06343325227499008,
-0.22282229363918304,
0.050492528825998306,
-0.08193226158618927,
0.0694160982966423,
-0.09974659234285355,
0.04703201726078987,
-0.019665813073515892,
0.07168370485305786,
-0.052492234855890274,
0.025007376447319984,
-0.023285027593374252,
0.0656980499625206,
0.029828233644366264,
0.1972961276769638,
-0.11733836680650711,
-0.0656147375702858,
0.05075822398066521,
-0.18322685360908508,
-0.2135290950536728,
0.04261644184589386,
-0.012876318767666817,
0.1303468942642212,
0.04075048491358757,
0.05000338703393936,
0.02827869914472103,
-0.06043403595685959,
0.0009076627320609987,
0.10918506234884262,
-0.06984000653028488,
-0.17179636657238007,
0.04915525019168854,
0.0642024576663971,
-0.1864931732416153,
0.04958309233188629,
0.0588674396276474,
0.10044735670089722,
-0.023717235773801804,
-0.07444756478071213,
-0.10238450020551682,
-0.08246321231126785,
-0.02005486562848091,
-0.04671313241124153,
-0.020074203610420227,
-0.03714090213179588,
0.005706062074750662,
0.1173885390162468,
0.13293230533599854,
-0.05285656452178955,
-0.00958924274891615,
-0.08579155802726746,
0.03487705439329147,
-0.11837325245141983,
0.05366849899291992,
-0.002173677086830139,
-0.030910659581422806,
-0.02025608904659748,
-0.010826515965163708,
-0.07775673270225525,
0.0786852315068245,
0.036689601838588715,
0.007802274543792009,
-0.04687562957406044,
-0.0036641538608819246,
0.10311242938041687,
0.07146940380334854,
-0.042056530714035034,
-0.1650473028421402,
-0.006933765951544046,
-0.0388154499232769,
0.18187232315540314,
-0.10326093435287476,
0.0768672451376915,
0.004125279374420643,
0.1317598521709442,
-0.058636005967855453,
0.06180243194103241,
0.014413111843168736,
-0.020901355892419815,
-0.06270238757133484,
0.008457247167825699,
0.07464863359928131,
0.031144024804234505,
-0.15256966650485992,
0.20289255678653717,
-0.13199731707572937,
0.14393596351146698,
0.18064405024051666,
-0.07909608632326126,
0.018092676997184753,
-0.02142081968486309,
0.02011336199939251,
-0.02660086564719677,
0.003718065330758691,
-0.0760800689458847,
-0.07703723758459091,
-0.046443190425634384,
0.0979800894856453,
-0.02609262987971306,
-0.030073028057813644,
0.04198761656880379,
-0.0012799074174836278,
-0.03210245072841644,
0.05373454466462135,
0.1493007391691208,
-0.1073472648859024,
0.14651088416576385,
0.15214349329471588,
0.02618745155632496,
0.032210785895586014,
-0.05961602181196213,
-0.03122686594724655,
-0.005816902033984661,
-0.03413346782326698,
-0.020621510222554207,
0.09718430042266846,
-0.05281573534011841,
0.06267273426055908,
0.11465589702129364,
-0.013615612871944904,
0.06530248373746872,
-0.09184747189283371,
-0.024764128029346466,
0.074228435754776,
-0.005236396566033363,
-0.01561465673148632,
0.10591375082731247,
-0.023863399401307106,
0.0496232844889164,
-0.04740089550614357,
-0.11022349447011948,
0.12373566627502441,
0.04035276174545288,
-0.117771215736866,
0.17479103803634644,
-0.0348825678229332,
-0.18626125156879425,
-0.14531321823596954,
-0.05859685316681862,
0.009744749404489994,
0.0056219566613435745,
0.0462927520275116,
0.03043273463845253,
-0.11163262277841568,
-0.10183367878198624,
0.022556845098733902,
-0.003898483933880925,
0.050566162914037704,
0.008603272028267384,
-0.0029876388143748045,
-0.04412614554166794,
-0.13829678297042847,
0.0006401907303370535,
0.08171524107456207,
0.05172149837017059,
0.10383907705545425,
-0.10390951484441757,
0.0795622393488884,
0.12724702060222626,
0.013787426985800266,
-0.025611044839024544,
0.026856601238250732,
0.13171452283859253,
-0.05604831501841545,
0.05153050273656845,
0.24687141180038452,
0.016771478578448296,
0.04014435037970543,
0.13695542514324188,
0.006555472034960985,
-0.060317911207675934,
0.04329945892095566,
-0.09760565310716629,
-0.11397405713796616,
-0.19032545387744904,
-0.13128967583179474,
-0.04741981253027916,
0.06949400156736374,
-0.018144313246011734,
0.08416276425123215,
0.11768069863319397,
0.17644579708576202,
-0.04715362936258316,
-0.025965815410017967,
-0.07908473163843155,
0.08845999836921692,
0.16905629634857178,
0.013473326340317726,
0.09700904786586761,
-0.15636040270328522,
-0.023721233010292053,
0.10331461578607559,
0.11282271891832352,
0.053974878042936325,
0.09524289518594742,
0.04293856769800186,
0.05771811306476593,
0.06069789081811905,
0.07102654874324799,
0.14547540247440338,
-0.0007726341718807817,
-0.061186786741018295,
-0.04188863933086395,
-0.03310469165444374,
0.04268127307295799,
0.061483707278966904,
-0.15084140002727509,
-0.017436835914850235,
-0.022162968292832375,
-0.0389728881418705,
0.043821487575769424,
0.0826880931854248,
0.04446735605597496,
-0.2816050350666046,
-0.017550546675920486,
0.10555847734212875,
-0.02307002991437912,
-0.04681321606040001,
-0.03715376928448677,
-0.0331912562251091,
-0.06599438935518265,
0.1701577752828598,
-0.014143829233944416,
0.06243368983268738,
-0.022465672343969345,
0.02525440603494644,
-0.029187925159931183,
0.009058787487447262,
0.037738073617219925,
0.08988083153963089,
-0.28683874011039734,
0.094550222158432,
0.0039205728098750114,
-0.013401882722973824,
-0.03568007051944733,
0.03558162972331047,
-0.02749280072748661,
0.2007301151752472,
0.011375023983418941,
-0.005865932907909155,
-0.0017795144813135266,
0.0438905693590641,
-0.03482551500201225,
0.07759219408035278,
-0.03997030854225159,
-0.07961104810237885,
0.01884450949728489,
-0.01850828155875206,
0.003535531461238861,
0.03154551982879639,
0.08391909301280975,
-0.09025971591472626,
-0.17337490618228912,
0.04585554823279381,
0.0951886922121048,
0.11409752815961838,
-0.035978689789772034,
-0.01927008666098118,
-0.2292359471321106,
0.1537909060716629,
-0.0660422146320343,
-0.1292884796857834,
-0.11764940619468689,
-0.13250480592250824,
0.016124563291668892,
0.0018255608156323433,
0.09318000078201294,
-0.01931716315448284,
0.037086185067892075,
-0.08131144195795059,
-0.19468311965465546,
0.11887994408607483,
-0.1508864015340805,
-0.0476420596241951,
0.010610656812787056,
0.037790726870298386,
-0.03710364177823067,
-0.007681541610509157,
0.02721276506781578,
-0.0019969684071838856,
-0.18665233254432678,
-0.10236801952123642,
-0.015822213143110275,
0.006794806569814682,
0.0060309963300824165,
0.013436248525977135,
0.049560390412807465,
-0.06913487613201141,
0.08233113586902618,
-0.11548716574907303,
0.0723276436328888,
0.18281321227550507,
-0.027577128261327744,
0.10674042999744415,
0.12911200523376465,
-0.024013936519622803,
-0.21522769331932068,
-0.11491293460130692,
-0.17089375853538513,
-0.0138707235455513,
0.004855223000049591,
-0.16043901443481445,
0.06159718334674835,
0.0623730793595314,
-0.06271310150623322,
0.011709854938089848,
-0.1952866166830063,
-0.10419099032878876,
0.15295816957950592,
0.0568196102976799,
0.3181999623775482,
-0.10362029075622559,
-0.056989289820194244,
-0.14145275950431824,
-0.2869182527065277,
0.17203712463378906,
-0.17805233597755432,
0.003911837004125118,
-0.0006680432707071304,
0.09999330341815948,
-0.006739133503288031,
-0.00927130039781332,
0.20087409019470215,
-0.037392787635326385,
0.03291580080986023,
-0.0795912966132164,
0.09695886820554733,
0.06124942749738693,
-0.009461194276809692,
0.13634613156318665,
-0.2443717122077942,
0.05021964758634567,
-0.09159889072179794,
-0.06368397921323776,
0.014489532448351383,
0.015060735866427422,
-0.013385681435465813,
-0.11171621084213257,
-0.05071209743618965,
-0.07854299247264862,
0.029256468638777733,
-0.026392711326479912,
0.1484663486480713,
0.0003113713173661381,
0.021248625591397285,
0.15554334223270416,
0.12136241048574448,
-0.04415322467684746,
0.04014987498521805,
-0.0663396343588829,
-0.10317439585924149,
0.06597846746444702,
-0.23527416586875916,
-0.0007714584353379905,
0.016475414857268333,
-0.07449447363615036,
0.002210434293374419,
0.06056447699666023,
0.02828073501586914,
-0.03434547409415245,
0.11781570315361023,
-0.13562224805355072,
-0.2162228673696518,
0.007467635907232761,
0.10526023805141449,
-0.026047643274068832,
0.17914599180221558,
0.1316654235124588,
-0.06380835920572281,
-0.03531791642308235,
0.019012194126844406,
0.03157501295208931,
-0.10256192833185196,
-0.009737396612763405,
0.02052437514066696,
0.07207799702882767,
-0.13399100303649902,
0.07333182543516159,
-0.035583656281232834,
-0.0640883594751358,
-0.02296130731701851,
0.02082696370780468,
-0.10767429322004318,
-0.10101567208766937,
0.0001650317426538095,
0.15173552930355072,
-0.2506720721721649,
-0.15680287778377533,
-0.09943879395723343,
-0.20939423143863678,
0.0582909919321537,
0.06572979688644409,
0.09974414110183716,
0.07614724338054657,
0.042050160467624664,
-0.022560514509677887,
-0.045745447278022766,
0.007964438758790493,
-0.06328336894512177,
0.11242825537919998,
-0.08771312981843948,
-0.06648494303226471,
-0.0689850002527237,
0.06838314980268478,
-0.0793171301484108,
0.016838179901242256,
-0.06357447803020477,
0.0033599920570850372,
-0.18346983194351196,
-0.06837128847837448,
-0.08002609759569168,
-0.028566021472215652,
0.027428293600678444,
-0.030017390847206116,
0.0014727552188560367,
0.026555584743618965,
-0.0404890775680542,
0.07800324261188507,
0.03453220799565315,
0.04528473690152168,
-0.030290337279438972,
0.022593434900045395,
0.012291481718420982,
-0.06730426847934723,
0.0870816707611084,
0.049456506967544556,
-0.0338517390191555,
-0.022980879992246628,
-0.14577369391918182,
0.04114202782511711,
0.08223826438188553,
0.03828035667538643,
-0.002453405177220702,
-0.060804396867752075,
-0.03282540664076805,
0.09539220482110977,
-0.06032995507121086,
0.019965089857578278,
0.0790373682975769,
-0.085635244846344,
0.06018412858247757,
-0.08641139417886734,
-0.019286826252937317,
-0.0514126792550087,
-0.07464037835597992,
0.02673937939107418,
0.03346014767885208,
0.12815791368484497,
-0.028271781280636787,
-0.05442959815263748,
-0.08120662719011307,
0.02681497111916542,
0.00495219137519598,
-0.10098660737276077,
-0.08363329619169235,
-0.03820842504501343,
-0.026264583691954613,
0.043463751673698425,
0.29362428188323975,
-0.08264505118131638,
-0.14799891412258148,
0.04363395273685455,
0.11366498470306396,
0.10091476887464523,
0.02365417778491974,
0.36137744784355164,
0.10000182688236237,
0.07344719022512436,
-0.13106614351272583,
0.051067475229501724,
0.06084422394633293,
-0.06801722943782806,
-0.039943087846040726,
0.08093162626028061,
-0.049585696309804916,
0.12802833318710327,
0.041060443967580795,
0.018373165279626846,
-0.008196541108191013,
-0.05138174444437027,
-0.058275315910577774,
-0.028868738561868668,
0.006988538894802332,
0.1435224562883377,
0.14991386234760284,
-0.026171645149588585,
0.008833928033709526,
0.07150404155254364,
0.00476393848657608,
-0.08617603033781052,
-0.12285705655813217,
-0.07085105776786804,
-0.18160535395145416,
-0.011142643168568611,
-0.02836593985557556,
-0.06484661251306534,
0.1382431983947754,
0.010882530361413956,
-0.02191752754151821,
-0.004287772811949253,
0.08488083630800247,
0.04841382056474686,
-0.0583573579788208,
-0.04381295666098595,
0.0022221438121050596,
-0.04310118779540062,
-0.1144879162311554,
-0.029188433662056923,
0.07151876389980316,
-0.033561695367097855,
0.023765547201037407,
0.07266220450401306,
0.08352877199649811,
-0.01812625676393509,
-0.07630458474159241,
-0.019504092633724213,
-0.006193626672029495,
-0.016396386548876762,
0.07419060170650482,
-0.021142875775694847,
0.0460834838449955,
0.05519969016313553,
0.10767647624015808,
-0.0740547850728035,
-0.12547583878040314,
-0.07977922260761261,
0.09873220324516296,
-0.10701785981655121,
0.10465975105762482,
0.09365487843751907,
-0.043788183480501175,
-0.12692612409591675,
0.18555407226085663,
0.21453046798706055,
-0.0011568487389013171,
0.02320794016122818,
-0.004683947190642357,
0.008888151496648788,
0.0379352830350399,
0.01750355213880539,
0.055394139140844345,
0.11387031525373459,
-0.06977062672376633,
0.049023356288671494,
-0.01929686777293682,
-0.028370190411806107,
-0.05312551558017731,
0.0960090160369873,
-0.10001430660486221,
-0.10983265191316605,
0.038153234869241714,
0.0664876401424408,
-0.10027481615543365,
0.020353222265839577,
-0.00974512193351984,
-0.01649296283721924,
-0.008959731087088585,
-0.10145766288042068,
0.11743849515914917,
-0.02669534459710121,
-0.013859155587852001,
-0.06359496712684631,
0.027553163468837738,
0.11801125109195709,
0.02157299965620041,
-0.1493699997663498,
-0.011534836143255234,
0.062050264328718185,
-0.0009420597343705595,
-0.020960578694939613,
0.002187091391533613,
0.08316365629434586,
0.08450585603713989,
0.02973143570125103,
-0.09096097201108932,
0.09365694969892502,
0.013019914738833904,
-0.03229924291372299,
0.08393039554357529,
-0.057395089417696,
-0.016788389533758163,
-0.02015235833823681,
-0.01676410809159279,
-0.12648151814937592,
-0.006624831352382898,
0.06027710437774658,
-0.08073479682207108,
-0.029026877135038376,
-0.023276951164007187,
-0.04539330303668976,
0.05938500910997391,
-0.006034390069544315,
-0.04889333248138428,
0.004982673563063145,
-0.00912618637084961,
0.08766312152147293,
0.011580334044992924,
-0.10180789977312088,
0.03149663284420967,
-0.06163351237773895,
-0.03820034861564636,
0.08207092434167862,
0.06812059879302979,
-0.17237988114356995,
-0.03934658691287041,
-0.15915171802043915,
-0.015087577514350414,
-0.1125597208738327,
-0.008747676387429237,
0.23046599328517914,
0.021560439839959145,
-0.052167028188705444,
-0.08178330212831497,
-0.013008864596486092,
0.059937965124845505,
-0.03571423888206482,
-0.09755008667707443
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-base-lora-1.77M-snli-model3
This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7590
- Accuracy: 0.7255
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 256
- eval_batch_size: 256
- seed: 84
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.5194 | 1.0 | 2146 | 0.4198 | 0.8470 |
| 0.4687 | 2.0 | 4292 | 0.3898 | 0.8574 |
| 0.4586 | 3.0 | 6438 | 0.3852 | 0.8599 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "t5-base", "model-index": [{"name": "t5-base-lora-1.77M-snli-model3", "results": []}]} | text-classification | varun-v-rao/t5-base-lora-1.77M-snli-model3 | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text-classification",
"generated_from_trainer",
"base_model:t5-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T07:08:12+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text-classification #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| t5-base-lora-1.77M-snli-model3
==============================
This model is a fine-tuned version of t5-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7590
* Accuracy: 0.7255
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 256
* eval\_batch\_size: 256
* seed: 84
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 84\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text-classification #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 84\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
74,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text-classification #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 84\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.08506516367197037,
0.09446289390325546,
-0.0021764938719570637,
0.11164454370737076,
0.14140236377716064,
0.0059529514983296394,
0.16060514748096466,
0.1175040453672409,
-0.07660192251205444,
0.04582127928733826,
0.12605127692222595,
0.12365572899580002,
0.02987685054540634,
0.13668763637542725,
-0.0862204059958458,
-0.21965326368808746,
0.013931778259575367,
0.02073846198618412,
-0.05457837134599686,
0.12793993949890137,
0.09696812927722931,
-0.11092774569988251,
0.09869305044412613,
-0.027079828083515167,
-0.14786545932292938,
0.009881940670311451,
0.023614414036273956,
-0.06756912171840668,
0.13523510098457336,
0.04750731214880943,
0.114339180290699,
0.02898639440536499,
0.06679307669401169,
-0.2051858901977539,
0.01369191613048315,
0.07040861994028091,
-0.012601260095834732,
0.08585821092128754,
0.042713381350040436,
0.00030625579529441893,
0.09404464811086655,
-0.10397406667470932,
0.053266655653715134,
0.02524605207145214,
-0.13331957161426544,
-0.18866227567195892,
-0.07099603116512299,
0.036306947469711304,
0.09954209625720978,
0.0866532251238823,
-0.017744116485118866,
0.1390954554080963,
-0.03350568562746048,
0.10548020154237747,
0.2089194506406784,
-0.3238099217414856,
-0.06084100902080536,
0.04401123896241188,
0.029810570180416107,
0.10833989083766937,
-0.09393820911645889,
-0.007147167343646288,
0.05874340608716011,
0.02846289612352848,
0.13738194108009338,
-0.024621164426207542,
-0.056857358664274216,
0.0007576591451652348,
-0.14342069625854492,
-0.03629192337393761,
0.17905084788799286,
0.06104148551821709,
-0.04568979889154434,
-0.05721690133213997,
-0.07876189798116684,
-0.14147743582725525,
-0.030859120190143585,
-0.004699500277638435,
0.047277677804231644,
-0.012319585308432579,
-0.04868658259510994,
-0.022054467350244522,
-0.10710066556930542,
-0.06530866026878357,
-0.05208112671971321,
0.11171002686023712,
0.03295687958598137,
-0.0004258729168213904,
-0.03178250417113304,
0.09505853801965714,
-0.016093699261546135,
-0.13775959610939026,
0.02146964706480503,
0.02412964217364788,
0.01024017482995987,
-0.0453575998544693,
-0.04425008222460747,
-0.0949132964015007,
0.02882090024650097,
0.11922905594110489,
-0.06719102710485458,
0.05982523411512375,
-0.018481262028217316,
0.04872865974903107,
-0.10560566931962967,
0.15807640552520752,
-0.03518212214112282,
-0.03342125564813614,
0.031926508992910385,
0.09002690017223358,
0.06707616150379181,
-0.0170166976749897,
-0.13437515497207642,
0.02529379166662693,
0.1251576691865921,
0.03670503944158554,
-0.04934750124812126,
0.08279820531606674,
-0.04497785121202469,
-0.00701766274869442,
0.03280971944332123,
-0.09429565817117691,
0.00835784338414669,
0.00014961037959437817,
-0.055519044399261475,
-0.07192391157150269,
0.0323820561170578,
0.021707642823457718,
-0.019588980823755264,
0.07963056862354279,
-0.07892928272485733,
-0.00027344137197360396,
-0.07322891056537628,
-0.1266668140888214,
0.021235182881355286,
-0.06874103099107742,
0.01143176294863224,
-0.12398239970207214,
-0.1775989830493927,
-0.006022178567945957,
0.05191720277070999,
-0.02510746754705906,
-0.04933478683233261,
-0.06093039736151695,
-0.0857648253440857,
0.019306126981973648,
-0.019086720421910286,
0.08593644946813583,
-0.07417982071638107,
0.0978156104683876,
0.03524143993854523,
0.05911409482359886,
-0.06169452145695686,
0.03432949632406235,
-0.1056232824921608,
0.03721662238240242,
-0.18350856006145477,
0.027162669226527214,
-0.03808470070362091,
0.07679684460163116,
-0.09714697301387787,
-0.07507093995809555,
0.00038632319774478674,
0.0000982890123850666,
0.06164129078388214,
0.09914211183786392,
-0.17154674232006073,
-0.04779786616563797,
0.17400310933589935,
-0.09013522416353226,
-0.1642010360956192,
0.1410110741853714,
-0.05123055726289749,
0.03524890914559364,
0.0703260526061058,
0.19871105253696442,
0.05750991776585579,
-0.07970836758613586,
0.012221438810229301,
0.0019987751729786396,
0.05998649075627327,
-0.04837323725223541,
0.08617924153804779,
-0.004349895752966404,
-0.016350798308849335,
0.006548846606165171,
-0.04828959330916405,
0.05525581166148186,
-0.06395486742258072,
-0.08163265883922577,
-0.0416206531226635,
-0.10133642703294754,
0.0394454151391983,
0.045711830258369446,
0.05636093020439148,
-0.12353493273258209,
-0.09155561774969101,
0.03873823583126068,
0.07115209847688675,
-0.07127275317907333,
0.02811557613313198,
-0.07197320461273193,
0.0813482329249382,
-0.059282973408699036,
-0.007320599630475044,
-0.15102706849575043,
-0.048204462975263596,
0.017640672624111176,
0.005379843525588512,
0.014559554867446423,
-0.010691083036363125,
0.06483213603496552,
0.07730003446340561,
-0.07016117870807648,
-0.03598930686712265,
0.001161162625066936,
0.015248006209731102,
-0.12308376282453537,
-0.20121237635612488,
-0.008881954476237297,
-0.025308024138212204,
0.1517964005470276,
-0.24315042793750763,
0.053768813610076904,
0.0038954918272793293,
0.078255295753479,
0.043128784745931625,
0.0004755939298775047,
-0.02108045481145382,
0.060730088502168655,
-0.053699422627687454,
-0.0655803456902504,
0.06520141661167145,
0.016865909099578857,
-0.10530149191617966,
-0.01510042604058981,
-0.17426370084285736,
0.18966007232666016,
0.1355343610048294,
-0.07842017710208893,
-0.0724993422627449,
0.0084282997995615,
-0.03582663834095001,
-0.026023348793387413,
-0.038181643933057785,
-0.0033732065930962563,
0.12880563735961914,
-0.006515000015497208,
0.16602392494678497,
-0.1003953143954277,
-0.0370006263256073,
0.02408105693757534,
-0.051025133579969406,
0.008812499232590199,
0.12864044308662415,
0.07945442199707031,
-0.14548887312412262,
0.15145151317119598,
0.18816445767879486,
-0.07541937381029129,
0.1420290470123291,
-0.048892825841903687,
-0.04958457872271538,
-0.021004559472203255,
0.03134926036000252,
0.021817289292812347,
0.09886457026004791,
-0.10380185395479202,
0.00538998655974865,
0.00975299533456564,
0.021661097183823586,
0.011891108937561512,
-0.21154646575450897,
-0.024927418678998947,
0.04172318056225777,
-0.06059065833687782,
0.0007147219148464501,
-0.023608779534697533,
-0.00771438330411911,
0.10936153680086136,
0.0029192743822932243,
-0.08648281544446945,
0.04758882895112038,
-0.008209743537008762,
-0.09361638873815536,
0.21077388525009155,
-0.08025410771369934,
-0.17793206870555878,
-0.13881616294384003,
-0.04368149861693382,
-0.057442523539066315,
0.027871036902070045,
0.06761484593153,
-0.07191573828458786,
-0.03894929215312004,
-0.13797783851623535,
-0.004591574426740408,
0.01703033037483692,
0.020198171958327293,
0.0011605002218857408,
0.008053048513829708,
0.08376239985227585,
-0.09812059998512268,
-0.013129675760865211,
-0.031002851203083992,
-0.045470062643289566,
0.031863074749708176,
0.008613544516265392,
0.11322059482336044,
0.15130019187927246,
-0.029010212048888206,
-0.001957297557964921,
-0.03969650715589523,
0.21628381311893463,
-0.061693523079156876,
0.002955944510176778,
0.13974259793758392,
-0.027085674926638603,
0.05309639126062393,
0.14385737478733063,
0.04619638994336128,
-0.09669037163257599,
0.03828758746385574,
0.02782745100557804,
-0.03547798469662666,
-0.22241389751434326,
-0.029167354106903076,
-0.04775240644812584,
0.026208985596895218,
0.08888521790504456,
0.03855331614613533,
0.03877386078238487,
0.05472341924905777,
0.01352967694401741,
0.0743362307548523,
0.0018316158093512058,
0.07431159168481827,
0.10790159553289413,
0.046976979821920395,
0.12525199353694916,
-0.06249347701668739,
-0.05294208601117134,
0.04160401225090027,
0.0017053033225238323,
0.18279118835926056,
0.017414992675185204,
0.16364718973636627,
0.04590875655412674,
0.15006177127361298,
0.006936248857527971,
0.06641922891139984,
-0.012577767483890057,
-0.030366238206624985,
-0.020313408225774765,
-0.05362468957901001,
-0.031368568539619446,
0.032131973654031754,
-0.0933929830789566,
0.05908321589231491,
-0.1103658527135849,
0.014354776591062546,
0.062306955456733704,
0.24008823931217194,
0.0516962930560112,
-0.3297589421272278,
-0.08497252315282822,
0.039174191653728485,
-0.02686612866818905,
-0.030494824051856995,
0.03684698045253754,
0.13466420769691467,
-0.057569436728954315,
0.04449226334691048,
-0.06773322820663452,
0.08888375759124756,
-0.04860943928360939,
0.048013195395469666,
0.047349438071250916,
0.07005935907363892,
-0.0208726916462183,
0.07503233850002289,
-0.29592499136924744,
0.2673482894897461,
0.015479412861168385,
0.06248147413134575,
-0.05397951230406761,
0.0007431514095515013,
0.02983100898563862,
0.09486972540616989,
0.07830154150724411,
-0.016658945009112358,
-0.08136985450983047,
-0.18204255402088165,
-0.06484297662973404,
0.0291590616106987,
0.08982621133327484,
-0.04354354739189148,
0.1047537550330162,
-0.04110012575984001,
0.00014409804134629667,
0.08538758009672165,
-0.002072729403153062,
-0.07518379390239716,
-0.0948110818862915,
-0.014808804728090763,
0.05579496547579765,
-0.02939537726342678,
-0.08309624344110489,
-0.08591488003730774,
-0.11659146100282669,
0.14242175221443176,
-0.06240755692124367,
-0.04427596554160118,
-0.09667685627937317,
0.048833608627319336,
0.04096753150224686,
-0.08309032022953033,
0.03544360771775246,
-0.0017955967923626304,
0.08858991414308548,
0.024629760533571243,
-0.0685187578201294,
0.13336493074893951,
-0.06860874593257904,
-0.1805010735988617,
-0.06247330456972122,
0.12071205675601959,
-0.0016164288390427828,
0.0425681434571743,
-0.003649151185527444,
0.00876554660499096,
-0.015264897607266903,
-0.07326310873031616,
0.0198354572057724,
-0.014256670139729977,
0.05998590216040611,
0.01816524751484394,
-0.05266280472278595,
-0.02074066363275051,
-0.06512869894504547,
-0.040215808898210526,
0.1730491667985916,
0.2863515019416809,
-0.07525868713855743,
0.0069122519344091415,
0.056904956698417664,
-0.06701602041721344,
-0.19528862833976746,
0.03216119855642319,
0.008160901255905628,
-0.005095175467431545,
0.04731419309973717,
-0.1397702544927597,
0.08632459491491318,
0.08530303835868835,
-0.028446728363633156,
0.10337164998054504,
-0.3055664300918579,
-0.13697023689746857,
0.12438680976629257,
0.15650738775730133,
0.12007511407136917,
-0.17615218460559845,
-0.04266125708818436,
-0.04953889176249504,
-0.09846699982881546,
0.12304779887199402,
-0.150518000125885,
0.1056756004691124,
-0.0025903754867613316,
0.03824999928474426,
0.007276549935340881,
-0.050431519746780396,
0.1244845986366272,
-0.023609977215528488,
0.11377055943012238,
-0.07848143577575684,
-0.008256315253674984,
0.05523382127285004,
-0.06455974280834198,
0.03547394648194313,
-0.15104050934314728,
0.04279359057545662,
-0.07190907001495361,
-0.026260480284690857,
-0.04626799374818802,
0.03204239159822464,
-0.038746077567338943,
-0.060023270547389984,
-0.04155610129237175,
0.022626522928476334,
0.05658453330397606,
-0.0036708477418869734,
0.17339180409908295,
0.0030771091114729643,
0.1568404585123062,
0.16179250180721283,
0.09054293483495712,
-0.048750437796115875,
-0.0034387512132525444,
0.0002644805354066193,
-0.037059612572193146,
0.05582214146852493,
-0.16823413968086243,
0.0395769402384758,
0.11594443023204803,
-0.0011287728557363153,
0.14842760562896729,
0.07102043181657791,
-0.03191258758306503,
0.0051165190525352955,
0.06653542816638947,
-0.1742372065782547,
-0.12955647706985474,
-0.021037491038441658,
-0.007830307818949223,
-0.11431646347045898,
0.05821377784013748,
0.13246603310108185,
-0.07704319804906845,
0.011174595914781094,
-0.006296055391430855,
0.027186967432498932,
-0.03086605854332447,
0.16097114980220795,
0.06761431694030762,
0.05379023775458336,
-0.07830886542797089,
0.0961032584309578,
0.06826198846101761,
-0.07523535192012787,
0.029381467029452324,
0.06619206070899963,
-0.09680785983800888,
-0.052422162145376205,
0.055825792253017426,
0.18877758085727692,
-0.025250541046261787,
-0.06787130236625671,
-0.1527402549982071,
-0.1262243241071701,
0.05427182838320732,
0.18857304751873016,
0.08590494096279144,
0.011964969336986542,
-0.024486757814884186,
-0.0016838439041748643,
-0.10954522341489792,
0.12656019628047943,
0.02928992547094822,
0.08232846111059189,
-0.1575172394514084,
0.12172000110149384,
-0.0011012919712811708,
0.012495900504291058,
-0.021505232900381088,
0.045848067849874496,
-0.11456756293773651,
0.0005024347337894142,
-0.12736985087394714,
0.009956765919923782,
-0.014858219772577286,
0.007597732357680798,
-0.0030867753084748983,
-0.04541231691837311,
-0.06025135889649391,
0.012075447477400303,
-0.09771779924631119,
-0.025140441954135895,
0.026902493089437485,
0.06086481735110283,
-0.12243131548166275,
-0.046723753213882446,
0.02149108611047268,
-0.07307111471891403,
0.0782104954123497,
0.002980064367875457,
0.006626482121646404,
0.055116329342126846,
-0.17053528130054474,
0.04080304875969887,
0.057425178587436676,
0.015387737192213535,
0.033780768513679504,
-0.08782657980918884,
-0.015542963519692421,
0.0009245594264939427,
0.03812411054968834,
0.024301636964082718,
0.09286097437143326,
-0.12761084735393524,
0.010166105814278126,
-0.01693710684776306,
-0.05426984652876854,
-0.052434395998716354,
0.0353119932115078,
0.06929697841405869,
0.0028506687376648188,
0.202727273106575,
-0.09784359484910965,
0.008376261219382286,
-0.20472490787506104,
0.007487709634006023,
0.003999542910605669,
-0.12783294916152954,
-0.12101530283689499,
-0.05077074095606804,
0.05429083853960037,
-0.06698653101921082,
0.12468444555997849,
0.022099336609244347,
0.03227929770946503,
0.04187237098813057,
-0.02681013010442257,
0.04964384809136391,
0.033821139484643936,
0.23237930238246918,
0.015797745436429977,
-0.04898327216506004,
0.02647770568728447,
0.015868723392486572,
0.11385130137205124,
0.082516610622406,
0.18510405719280243,
0.17406918108463287,
-0.04259245842695236,
0.10235702246427536,
0.03660227730870247,
-0.039178408682346344,
-0.12907080352306366,
0.05414614826440811,
-0.03862419351935387,
0.11892339587211609,
-0.01669742912054062,
0.20891547203063965,
0.11124338954687119,
-0.14834369719028473,
0.009517048485577106,
-0.06792688369750977,
-0.07799404114484787,
-0.11290126293897629,
-0.07372262328863144,
-0.09971234202384949,
-0.13742142915725708,
-0.01800502836704254,
-0.11647263169288635,
0.013367044739425182,
0.11804603040218353,
0.005421522539108992,
-0.023597251623868942,
0.16457466781139374,
0.016482753679156303,
0.01637876220047474,
0.06652171909809113,
-0.0010463747894391418,
-0.03556444123387337,
-0.07338521629571915,
-0.09251993894577026,
0.015425246208906174,
-0.022530335932970047,
0.021069476380944252,
-0.03576383367180824,
-0.021539321169257164,
0.0486154779791832,
-0.018533028662204742,
-0.10717687010765076,
0.01377134770154953,
0.026589734479784966,
0.04962271824479103,
0.06008307635784149,
0.023230522871017456,
-0.002863008063286543,
0.010776408948004246,
0.23829397559165955,
-0.07650849223136902,
-0.07289448380470276,
-0.08887667208909988,
0.21882130205631256,
0.02111099101603031,
0.0026692855171859264,
0.010676548816263676,
-0.0957178920507431,
0.026498762890696526,
0.21854737401008606,
0.18890133500099182,
-0.08733202517032623,
-0.0003200331993866712,
-0.028728527948260307,
-0.005071296356618404,
-0.02344394475221634,
0.10963720828294754,
0.1201392188668251,
0.01416444405913353,
-0.07550885528326035,
-0.02882133051753044,
-0.03649395704269409,
0.004879088141024113,
-0.04917005077004433,
0.07387475669384003,
0.03205997869372368,
0.009457645937800407,
-0.04036645591259003,
0.06911344826221466,
-0.04898340255022049,
-0.1147141084074974,
0.01098004449158907,
-0.1993015706539154,
-0.144868403673172,
-0.02456207200884819,
0.12060631811618805,
-0.012634440325200558,
0.043798722326755524,
-0.03318054601550102,
0.00902690552175045,
0.04197103902697563,
-0.0202048160135746,
-0.06810525804758072,
-0.06350395083427429,
0.057246290147304535,
-0.1229688823223114,
0.22029197216033936,
-0.03553556650876999,
0.036325350403785706,
0.1336647868156433,
0.030686907470226288,
-0.07886232435703278,
0.10564736276865005,
0.048493850976228714,
-0.03543952852487564,
0.04549747332930565,
0.07975666970014572,
-0.03674027696251869,
0.11948411166667938,
0.06263840198516846,
-0.12154913693666458,
0.007761893328279257,
-0.049306996166706085,
-0.07656164467334747,
-0.058587998151779175,
-0.04088728502392769,
-0.0648215264081955,
0.13066814839839935,
0.17589205503463745,
-0.04879168048501015,
0.01266532950103283,
-0.050702668726444244,
0.006776174530386925,
0.062221843749284744,
0.019378671422600746,
-0.022961225360631943,
-0.22301094233989716,
0.019161269068717957,
0.06780354678630829,
0.005786904599517584,
-0.3149506151676178,
-0.0714186504483223,
-0.022714484483003616,
-0.03781113028526306,
-0.10933323949575424,
0.08965089917182922,
0.1347602903842926,
0.04607947915792465,
-0.06786313652992249,
-0.0721641406416893,
-0.07214333862066269,
0.16681762039661407,
-0.1184430941939354,
-0.10115151852369308
] |
null | null | null | https://civitai.com/models/184528?modelVersionId=207142 | {} | null | Yuriy81/VasilyLozhkin_lora | [
"region:us"
] | 2024-02-06T07:09:25+00:00 | [] | [] | TAGS
#region-us
| URL | [] | [
"TAGS\n#region-us \n"
] | [
6
] | [
"passage: TAGS\n#region-us \n"
] | [
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# wkqco33/transformer_study
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.7143
- Validation Loss: 0.5352
- Train Accuracy: 0.9
- Epoch: 2
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 3e-05, 'decay_steps': 12000, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Train Accuracy | Epoch |
|:----------:|:---------------:|:--------------:|:-----:|
| 2.7907 | 1.6296 | 0.816 | 0 |
| 1.2019 | 0.8294 | 0.883 | 1 |
| 0.7143 | 0.5352 | 0.9 | 2 |
### Framework versions
- Transformers 4.35.2
- TensorFlow 2.15.0
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_keras_callback"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "wkqco33/transformer_study", "results": []}]} | image-classification | wkqco33/transformer_study | [
"transformers",
"tf",
"vit",
"image-classification",
"generated_from_keras_callback",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T07:12:44+00:00 | [] | [] | TAGS
#transformers #tf #vit #image-classification #generated_from_keras_callback #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| wkqco33/transformer\_study
==========================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on an unknown dataset.
It achieves the following results on the evaluation set:
* Train Loss: 0.7143
* Validation Loss: 0.5352
* Train Accuracy: 0.9
* Epoch: 2
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* optimizer: {'name': 'AdamWeightDecay', 'learning\_rate': {'module': 'keras.optimizers.schedules', 'class\_name': 'PolynomialDecay', 'config': {'initial\_learning\_rate': 3e-05, 'decay\_steps': 12000, 'end\_learning\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\_name': None}, 'decay': 0.0, 'beta\_1': 0.9, 'beta\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\_decay\_rate': 0.01}
* training\_precision: float32
### Training results
### Framework versions
* Transformers 4.35.2
* TensorFlow 2.15.0
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 3e-05, 'decay\\_steps': 12000, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: float32",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tf #vit #image-classification #generated_from_keras_callback #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 3e-05, 'decay\\_steps': 12000, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: float32",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
73,
227,
4,
31
] | [
"passage: TAGS\n#transformers #tf #vit #image-classification #generated_from_keras_callback #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* optimizer: {'name': 'AdamWeightDecay', 'learning\\_rate': {'module': 'keras.optimizers.schedules', 'class\\_name': 'PolynomialDecay', 'config': {'initial\\_learning\\_rate': 3e-05, 'decay\\_steps': 12000, 'end\\_learning\\_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered\\_name': None}, 'decay': 0.0, 'beta\\_1': 0.9, 'beta\\_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight\\_decay\\_rate': 0.01}\n* training\\_precision: float32### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* TensorFlow 2.15.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.06904222071170807,
0.09147489815950394,
-0.007227006834000349,
0.08273784071207047,
0.14767566323280334,
0.0562569722533226,
0.1297539323568344,
0.123295359313488,
-0.08313818275928497,
0.13441653549671173,
0.09632871299982071,
0.10123317688703537,
0.059004731476306915,
0.11456355452537537,
-0.05947503447532654,
-0.14696024358272552,
0.03985626995563507,
-0.04118388891220093,
-0.05819741636514664,
0.06719490885734558,
0.07495816797018051,
-0.06959211826324463,
0.0900619700551033,
-0.03529242053627968,
-0.09428901970386505,
0.00019961825455538929,
0.031248755753040314,
-0.049872227013111115,
0.08873815834522247,
0.06327314674854279,
0.07895998656749725,
0.02258102409541607,
0.01645777001976967,
-0.1961069256067276,
0.0013896849704906344,
0.11994800716638565,
0.003038331400603056,
0.07427183538675308,
0.04777451604604721,
-0.044153373688459396,
0.0986330583691597,
-0.09340760856866837,
0.049577757716178894,
0.03998931124806404,
-0.1421799510717392,
-0.23361358046531677,
-0.07745612412691116,
0.0299662034958601,
0.07941281050443649,
0.07355644553899765,
0.01078666653484106,
0.10267055034637451,
-0.08116647601127625,
0.08117882162332535,
0.12472599744796753,
-0.22980698943138123,
-0.05243504047393799,
0.06849102675914764,
-0.01728690043091774,
0.03422502428293228,
-0.07786723226308823,
0.0012834483059123158,
0.0009237852646037936,
0.012008788995444775,
0.032121967524290085,
0.005153452046215534,
-0.03761964291334152,
-0.04376654699444771,
-0.04503650590777397,
-0.051079437136650085,
0.12843537330627441,
0.06362666934728622,
-0.02883228287100792,
-0.053701914846897125,
-0.029143119230866432,
-0.19778572022914886,
-0.026355836540460587,
0.00005705314106307924,
0.034094344824552536,
0.01741166226565838,
-0.039494674652814865,
-0.00029822494252584875,
-0.042419735342264175,
-0.04530969634652138,
0.02359204739332199,
0.061080195009708405,
0.028975438326597214,
0.03467737138271332,
0.004277931991964579,
0.05191458389163017,
-0.03417497128248215,
-0.11829128116369247,
-0.03152389079332352,
0.0005137814441695809,
-0.061638131737709045,
-0.02866506762802601,
-0.05092477798461914,
-0.006045190617442131,
0.09876039624214172,
0.15967613458633423,
-0.07051581889390945,
0.12722782790660858,
-0.014868508093059063,
0.03877470642328262,
-0.10642540454864502,
0.08889195322990417,
0.0028793623205274343,
-0.027127793058753014,
-0.007814239710569382,
0.06641735136508942,
0.000786525895819068,
-0.037479665130376816,
-0.05285016447305679,
0.03869781643152237,
0.09535927325487137,
0.02614372782409191,
-0.009771516546607018,
0.09887414425611496,
-0.078006811439991,
-0.006109306123107672,
0.006062102038413286,
-0.10242995619773865,
0.04564454406499863,
0.05581618472933769,
-0.0798151046037674,
0.04653175175189972,
0.079592265188694,
-0.008899522945284843,
-0.05799827352166176,
0.05110582709312439,
-0.059006597846746445,
-0.01594051532447338,
-0.0938362181186676,
-0.08909279853105545,
0.028175026178359985,
-0.07370617240667343,
-0.03469507396221161,
-0.08494420349597931,
-0.14482621848583221,
-0.07845780998468399,
0.10340604186058044,
-0.05100495368242264,
-0.0329294353723526,
-0.07125435769557953,
-0.1515665203332901,
0.057175181806087494,
0.0007460592896677554,
0.10458886623382568,
-0.0601198673248291,
0.06527891755104065,
-0.016280174255371094,
0.043786242604255676,
0.007391185499727726,
0.03338747099041939,
-0.05771936476230621,
0.038410987704992294,
-0.16963961720466614,
0.1198216900229454,
-0.09541516751050949,
0.07630891352891922,
-0.1534680426120758,
-0.05521896108984947,
0.023273605853319168,
0.012159116566181183,
0.0960538387298584,
0.11505968868732452,
-0.1602262705564499,
-0.0620303750038147,
0.120844267308712,
-0.09037720412015915,
-0.08046112209558487,
0.06524492800235748,
-0.02299126610159874,
-0.02321157231926918,
0.07837678492069244,
0.07670532166957855,
0.07091910392045975,
-0.08345920592546463,
0.018479548394680023,
-0.06595604866743088,
0.033056702464818954,
0.032172080129384995,
0.029990747570991516,
-0.07451392710208893,
-0.09196655452251434,
0.037902504205703735,
-0.01907285861670971,
0.002539119916036725,
-0.06119970232248306,
-0.06315837800502777,
-0.03764403983950615,
-0.06796081364154816,
0.024765759706497192,
0.04166458174586296,
0.018578888848423958,
-0.07806036621332169,
-0.17003831267356873,
0.04746958240866661,
0.04386327415704727,
-0.06767246872186661,
0.02154446765780449,
-0.06322609633207321,
0.06128772348165512,
0.05318376049399376,
0.00041060245712287724,
-0.15693670511245728,
-0.10307876765727997,
0.023742295801639557,
-0.025619857013225555,
0.028482122346758842,
-0.054863039404153824,
0.05997494235634804,
0.035463228821754456,
-0.05819549784064293,
-0.0032069035805761814,
-0.0041874428279697895,
0.01188791822642088,
-0.04201396554708481,
-0.2332635074853897,
-0.02958635427057743,
-0.005634973291307688,
0.10737644135951996,
-0.28163087368011475,
0.003394098486751318,
0.06889469176530838,
0.12964321672916412,
0.04129800200462341,
-0.04355150833725929,
-0.034336186945438385,
0.059306032955646515,
-0.01849990151822567,
-0.07650992274284363,
0.04927570000290871,
0.014202234335243702,
-0.09759939461946487,
-0.06529154628515244,
-0.14665761590003967,
0.07600293308496475,
0.11652173846960068,
-0.0941689983010292,
-0.1461714208126068,
0.018208656460046768,
-0.02595602162182331,
-0.035215042531490326,
0.0032825672533363104,
0.035934340208768845,
0.13672518730163574,
0.033635158091783524,
0.12974907457828522,
-0.0319262333214283,
-0.004498585592955351,
0.014372939243912697,
-0.013133767992258072,
-0.01903371699154377,
0.11989094316959381,
0.040921274572610855,
-0.0630682110786438,
0.09032420814037323,
0.0439426526427269,
-0.12974417209625244,
0.09727347642183304,
-0.055518992245197296,
-0.04912499338388443,
-0.06244800239801407,
0.06067074462771416,
0.058717530220746994,
0.04609327018260956,
-0.11273910105228424,
-0.0020142432767897844,
0.015457076951861382,
-0.00018496488337405026,
-0.0010310497600585222,
-0.14371120929718018,
0.024319931864738464,
-0.009221025742590427,
-0.05782395228743553,
0.07014104723930359,
-0.020361142233014107,
0.00927723478525877,
0.10328429937362671,
0.03851591795682907,
-0.03701514005661011,
0.04914304241538048,
-0.023482447490096092,
-0.07748202979564667,
0.2070673108100891,
-0.10857846587896347,
-0.12792640924453735,
-0.10181865841150284,
-0.010166577063500881,
-0.0633077323436737,
-0.012769371271133423,
0.001467878115363419,
-0.08551733940839767,
-0.07298566401004791,
-0.06810281425714493,
-0.03719024360179901,
-0.023116225376725197,
0.005058037582784891,
-0.00976504385471344,
0.03947959095239639,
0.14365418255329132,
-0.08758646994829178,
-0.03146136552095413,
0.0003046812489628792,
-0.09655066579580307,
0.012112904340028763,
0.018276715651154518,
-0.00028042058693245053,
0.12184596061706543,
-0.002968926215544343,
0.01744452863931656,
-0.03301549702882767,
0.21056655049324036,
-0.06049687787890434,
0.02861015312373638,
0.1174691766500473,
0.00016398608568124473,
0.07344882935285568,
0.17056234180927277,
0.0645790547132492,
-0.09736859798431396,
0.031193841248750687,
0.08613468706607819,
-0.016485806554555893,
-0.22738593816757202,
-0.028556115925312042,
-0.045212917029857635,
-0.0748375952243805,
0.091580830514431,
0.06237906217575073,
0.16908732056617737,
0.02620084024965763,
0.002995668677613139,
0.07794608175754547,
0.06445538997650146,
0.08883446455001831,
0.13341712951660156,
0.09865345805883408,
0.10810176283121109,
-0.031029112637043,
0.030743146315217018,
0.03084629215300083,
-0.013071773573756218,
0.20459677278995514,
0.013520627282559872,
0.08699826896190643,
0.09432916343212128,
0.06997895240783691,
0.014480089768767357,
-0.037744127213954926,
0.00958859734237194,
0.01689353585243225,
0.022264746949076653,
-0.07568367570638657,
-0.04771256819367409,
0.04594780132174492,
0.03801079839468002,
0.038084354251623154,
-0.08043257892131805,
0.005333815701305866,
0.05860786512494087,
0.23168030381202698,
0.10747697949409485,
-0.32324835658073425,
-0.09802089631557465,
0.005640728399157524,
-0.009994414635002613,
-0.04875531047582626,
-0.005818197038024664,
0.031220806762576103,
-0.07233339548110962,
0.09497153013944626,
-0.04275177791714668,
0.07516129314899445,
-0.07882103323936462,
0.03793886676430702,
0.1286989003419876,
0.10490463674068451,
0.024709532037377357,
0.010155831463634968,
-0.3406110107898712,
0.26025518774986267,
0.012593073770403862,
0.11499462276697159,
-0.03765209764242172,
0.059336427599191666,
0.04307185485959053,
-0.017872042953968048,
0.06419318914413452,
-0.012806049548089504,
-0.13914857804775238,
-0.17120231688022614,
-0.05656205490231514,
-0.004743640311062336,
0.10653779655694962,
-0.04935711249709129,
0.08003420382738113,
-0.037348076701164246,
-0.017778033390641212,
0.039616119116544724,
-0.005997066386044025,
-0.1870489865541458,
-0.07824354618787766,
0.05377478152513504,
0.0324053056538105,
0.02250288985669613,
-0.06363028287887573,
-0.05761082470417023,
-0.09757882356643677,
0.2177230417728424,
-0.13941830396652222,
-0.05714932084083557,
-0.13963374495506287,
0.08058393001556396,
0.09927540272474289,
-0.0587504543364048,
0.0551971010863781,
-0.028897877782583237,
0.07818105816841125,
0.07142694294452667,
-0.07248037308454514,
0.12566179037094116,
-0.009285001084208488,
-0.21157628297805786,
-0.07668408751487732,
0.10408537089824677,
0.026237428188323975,
0.014100317843258381,
-0.018600475043058395,
0.07854540646076202,
0.042514339089393616,
-0.08176938444375992,
0.08267394453287125,
0.05568770319223404,
0.055755864828825,
0.04932232201099396,
-0.039648786187171936,
-0.06539496779441833,
-0.03651726245880127,
0.0037619154900312424,
0.05573459342122078,
0.28939107060432434,
-0.08415021747350693,
0.04149454087018967,
0.04282255470752716,
-0.09745068848133087,
-0.1872415542602539,
0.05978880822658539,
0.10976250469684601,
-0.016893813386559486,
-0.08053860813379288,
-0.21056810021400452,
0.07761231064796448,
0.10782158374786377,
-0.019436465576291084,
0.056913819164037704,
-0.263360857963562,
-0.15117037296295166,
0.05280141904950142,
0.10230211168527603,
-0.0007788894581608474,
-0.18487471342086792,
-0.07495270669460297,
-0.0740542784333229,
-0.06787767261266708,
0.1418408751487732,
-0.037831127643585205,
0.08946246653795242,
0.024401502683758736,
-0.00678542023524642,
0.022034918889403343,
-0.03245911747217178,
0.1434483379125595,
-0.0026696613058447838,
0.08857452124357224,
-0.054507065564394,
-0.024459682404994965,
0.07200295478105545,
-0.1029820442199707,
0.026019470766186714,
-0.04678291454911232,
0.036840539425611496,
-0.11066929250955582,
0.0016805361956357956,
-0.07044815272092819,
0.06251175701618195,
-0.06495136767625809,
0.0014767268439754844,
-0.027611149474978447,
0.06965743005275726,
0.08634661138057709,
0.012137116864323616,
0.12180393189191818,
-0.04285716265439987,
0.198931485414505,
0.14569614827632904,
0.07018261402845383,
0.017167020589113235,
-0.07524346560239792,
0.05360113084316254,
-0.03505532443523407,
0.06583015620708466,
-0.17755313217639923,
0.0571967288851738,
0.12982726097106934,
-0.0008873171755112708,
0.13002946972846985,
0.06001335009932518,
-0.0406407006084919,
0.008349095471203327,
0.07051368802785873,
-0.09150835871696472,
-0.05592271313071251,
0.01434549130499363,
-0.002217108616605401,
-0.06062610074877739,
0.014591491781175137,
0.1426590085029602,
-0.038029927760362625,
0.02820228971540928,
0.024455571547150612,
0.04912155121564865,
-0.05347076430916786,
0.09661110490560532,
0.015117775648832321,
0.08752407878637314,
-0.08218062669038773,
0.13744114339351654,
0.10509933531284332,
-0.1108546257019043,
0.0874013751745224,
0.061357978731393814,
-0.07287874817848206,
-0.03796515613794327,
0.05580111965537071,
0.12128829956054688,
0.05845325067639351,
-0.043858665972948074,
-0.07107144594192505,
-0.1424241065979004,
0.09133335947990417,
0.16389766335487366,
0.01598849892616272,
0.057489775121212006,
-0.01098124124109745,
0.0032169006299227476,
-0.09172666072845459,
0.060887571424245834,
0.03592981398105621,
0.05501928925514221,
-0.12603089213371277,
0.15015050768852234,
0.014740919694304466,
-0.03823891654610634,
0.0032935920171439648,
0.0015600149054080248,
-0.1986069679260254,
-0.0029087269213050604,
-0.11050613224506378,
0.05074112117290497,
0.01403084397315979,
0.007499920669943094,
0.04130756855010986,
-0.037667326629161835,
-0.05772634223103523,
0.022426843643188477,
-0.09918785840272903,
-0.06085958331823349,
0.04722100496292114,
0.08706890046596527,
-0.11870910227298737,
-0.04968101531267166,
0.014830164611339569,
-0.11562264710664749,
0.048183124512434006,
0.01294796820729971,
-0.003927813842892647,
0.014010941609740257,
-0.1283210963010788,
0.019327323883771896,
0.02191939763724804,
0.002976374700665474,
0.010922503657639027,
-0.12010970711708069,
0.023162193596363068,
-0.0342065766453743,
0.03151071444153786,
0.01563703641295433,
0.059715379029512405,
-0.09440819919109344,
-0.04160333797335625,
-0.0307682566344738,
-0.031275056302547455,
-0.02913776785135269,
0.04928196221590042,
0.15024474263191223,
-0.036126360297203064,
0.15431371331214905,
-0.11463943868875504,
0.03232420235872269,
-0.1901809275150299,
-0.007565273903310299,
0.018255027011036873,
-0.06734558194875717,
-0.1199609562754631,
-0.028144050389528275,
0.1207488626241684,
-0.09529861062765121,
0.09098049998283386,
0.00022882487974129617,
0.08460511267185211,
0.029041705653071404,
-0.07249404489994049,
-0.09150657057762146,
0.0932013988494873,
0.16351617872714996,
0.07932014763355255,
0.0004088583227712661,
0.08940811455249786,
-0.034879542887210846,
0.04630595073103905,
0.05916230380535126,
0.17542915046215057,
0.1336033046245575,
0.010576733388006687,
0.07348114252090454,
0.06061619520187378,
-0.10191037505865097,
-0.08776376396417618,
0.18057724833488464,
-0.09611358493566513,
0.17507359385490417,
-0.08515521138906479,
0.07609785348176956,
0.028945699334144592,
-0.16127124428749084,
0.046412356197834015,
-0.07517636567354202,
-0.08892756700515747,
-0.09995808452367783,
-0.13374246656894684,
-0.10358821600675583,
-0.11020481586456299,
0.005266417283564806,
-0.08605042845010757,
0.007905213162302971,
0.11014051735401154,
0.023653924465179443,
0.01511160098016262,
0.03931911662220955,
-0.045137666165828705,
0.025386648252606392,
0.11399489641189575,
-0.006980685982853174,
-0.020118940621614456,
-0.05911397561430931,
-0.07398420572280884,
0.04114735871553421,
0.023155895993113518,
0.028969749808311462,
0.021670149639248848,
0.00034607481211423874,
0.06084809452295303,
0.010886432603001595,
-0.09971659630537033,
0.07523790001869202,
0.008492069318890572,
-0.013120634481310844,
0.06917154043912888,
0.022645164281129837,
-0.019782599061727524,
-0.009201325476169586,
0.1385904848575592,
-0.07741670310497284,
-0.05802226811647415,
-0.15338172018527985,
0.24537932872772217,
-0.03188189119100571,
0.022369764745235443,
0.000299682782497257,
-0.07138606905937195,
-0.020526953041553497,
0.1637636125087738,
0.15209518373012543,
-0.04113563522696495,
-0.020312657579779625,
0.09080825001001358,
-0.01998049207031727,
-0.039859041571617126,
0.13809292018413544,
0.05135064199566841,
-0.029860442504286766,
-0.038702432066202164,
-0.02733994461596012,
0.014348400756716728,
-0.03172214329242706,
-0.0767325833439827,
0.07618486881256104,
-0.018700480461120605,
-0.016158955171704292,
-0.03186481446027756,
0.06812075525522232,
-0.11389068514108658,
-0.1116330549120903,
0.15312719345092773,
-0.20131845772266388,
-0.17656196653842926,
-0.02680782973766327,
0.015704559162259102,
0.017307836562395096,
0.024327648803591728,
-0.007775465026497841,
-0.026974622160196304,
0.11678119748830795,
-0.05481180548667908,
-0.01523995865136385,
-0.10923726111650467,
0.012837016955018044,
-0.015198552049696445,
0.22682417929172516,
-0.024348661303520203,
0.029541613534092903,
0.147271066904068,
0.020914118736982346,
-0.08861443400382996,
0.042771998792886734,
0.0733485147356987,
-0.11689503490924835,
0.039727333933115005,
0.0794232115149498,
-0.030030453577637672,
0.16799403727054596,
0.09019970893859863,
-0.08583082258701324,
0.011730344034731388,
-0.01295433659106493,
-0.061678197234869,
-0.0455196276307106,
-0.042692799121141434,
-0.07417385280132294,
0.11763536930084229,
0.21686409413814545,
-0.03021974302828312,
-0.020997995510697365,
-0.041259732097387314,
0.038673706352710724,
0.023528490215539932,
0.019950417801737785,
-0.07842183858156204,
-0.21019858121871948,
0.08137401938438416,
0.022302234545350075,
0.07042843848466873,
-0.13907334208488464,
-0.07669954746961594,
0.020510347560048103,
-0.007680200971662998,
-0.10447739064693451,
0.1165570542216301,
0.06257680058479309,
0.030228929594159126,
-0.056864649057388306,
-0.15483437478542328,
-0.017817383632063866,
0.18638354539871216,
-0.11415910720825195,
-0.07248587161302567
] |
null | null | transformers | # miqu-1-120b

* EXL2: [2.4bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2) | [2.65bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2) | [3.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2) | [4.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2) | 5.0bpw
* GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/) | [IQ3_XXS](https://huggingface.co/wolfram/miqu-1-120b-GGUF)
* HF FP16: [wolfram/miqu-1-120b](https://huggingface.co/wolfram/miqu-1-120b)
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
Inspired by [Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2), [MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [goliath-120b](https://huggingface.co/alpindale/goliath-120b).
Thanks for the support, [CopilotKit](https://github.com/CopilotKit/CopilotKit) - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, [Lone Striker](https://huggingface.co/LoneStriker)!
## Prompt template: Mistral
```
<s>[INST] {prompt} [/INST]
```
See also: [🐺🐦⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates : LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/18ljvxb/llm_prompt_format_comparisontest_mixtral_8x7b/)
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
dtype: float16
merge_method: passthrough
slices:
- sources:
- layer_range: [0, 20]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [10, 30]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [20, 40]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [30, 50]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [40, 60]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [50, 70]
model: 152334H/miqu-1-70b-sf
- sources:
- layer_range: [60, 80]
model: 152334H/miqu-1-70b-sf
```
## Credits & Special Thanks
* original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
* leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
* f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
* mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
* mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
### Support
* [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| {"language": ["en", "de", "fr", "es", "it"], "library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["152334H/miqu-1-70b-sf"]} | text-generation | LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"en",
"de",
"fr",
"es",
"it",
"base_model:152334H/miqu-1-70b-sf",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T07:12:47+00:00 | [] | [
"en",
"de",
"fr",
"es",
"it"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # miqu-1-120b
!image/jpeg
* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw
* GGUF: Q2_K-Q5_K_M | IQ3_XXS
* HF FP16: wolfram/miqu-1-120b
This is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.
Inspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.
Thanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
Thanks for the EXL2 and GGUF quants, Lone Striker!
## Prompt template: Mistral
See also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA
## Model Details
* Max Context: 32764 tokens (kept the weird number from the original/base model)
* Layers: 140
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* 152334H/miqu-1-70b-sf
### Configuration
The following YAML configuration was used to produce this model:
## Credits & Special Thanks
* original (unreleased) model: mistralai (Mistral AI_)
* leaked model: miqudev/miqu-1-70b
* f16 model: 152334H/miqu-1-70b-sf
* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.
* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2
### Support
* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
| [
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!",
"## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA",
"## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Credits & Special Thanks\n\n* original (unreleased) model: mistralai (Mistral AI_)\n* leaked model: miqudev/miqu-1-70b\n* f16 model: 152334H/miqu-1-70b-sf\n* mergekit: arcee-ai/mergekit: Tools for merging pretrained large language models.\n* mergekit_config.yml: nsfwthrowitaway69/Venus-120b-v1.2",
"### Support\n\n* My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!",
"#### DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK."
] | [
85,
206,
44,
31,
4,
17,
28,
17,
107,
69,
43
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #conversational #en #de #fr #es #it #base_model-152334H/miqu-1-70b-sf #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# miqu-1-120b\n\n!image/jpeg\n\n* EXL2: 2.4bpw | 2.65bpw | 3.0bpw | 4.0bpw | 5.0bpw\n* GGUF: Q2_K-Q5_K_M | IQ3_XXS\n* HF FP16: wolfram/miqu-1-120b\n\nThis is a 120b frankenmerge of miqu-1-70b created by interleaving layers of miqu-1-70b-sf with itself using mergekit.\n\nInspired by Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b.\n\nThanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.\n\nThanks for the EXL2 and GGUF quants, Lone Striker!## Prompt template: Mistral\n\n\n\nSee also: ⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA## Model Details\n\n* Max Context: 32764 tokens (kept the weird number from the original/base model)\n* Layers: 140## Merge Details### Merge Method\n\nThis model was merged using the passthrough merge method.### Models Merged\n\nThe following models were included in the merge:\n* 152334H/miqu-1-70b-sf### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.08452296257019043,
0.07996358722448349,
-0.006216804962605238,
0.07489742338657379,
0.062447689473629,
0.048032015562057495,
0.10850238054990768,
0.11730548739433289,
0.04997159540653229,
0.1304655373096466,
0.019059622660279274,
0.0882313996553421,
0.10314993560314178,
0.1546645164489746,
-0.0008784975507296622,
-0.13700617849826813,
0.03406155854463577,
-0.032812073826789856,
-0.05265757441520691,
0.07489487528800964,
0.07002948969602585,
-0.049538981169462204,
0.07827795296907425,
0.01992727816104889,
-0.0794241651892662,
-0.02351442351937294,
-0.03350690379738808,
-0.0034194772597402334,
0.06382977962493896,
0.09272785484790802,
0.010276054963469505,
0.01770416460931301,
0.010751184076070786,
-0.1683615893125534,
0.032938648015260696,
0.03592948243021965,
-0.017932593822479248,
0.07718798518180847,
0.10788709670305252,
-0.018908003345131874,
0.021088771522045135,
-0.10027661174535751,
-0.02568136155605316,
0.08249933272600174,
-0.10719147324562073,
-0.1474289745092392,
-0.18762712180614471,
0.15491966903209686,
0.042507048696279526,
0.03724713996052742,
-0.011902958154678345,
0.1333865076303482,
0.050249699503183365,
0.06016452983021736,
0.24593202769756317,
-0.22131793200969696,
-0.061919860541820526,
0.028957664966583252,
0.05541115999221802,
-0.04985284432768822,
-0.021356018260121346,
0.01992829702794552,
0.04054830223321915,
0.02191905304789543,
-0.020568734034895897,
-0.04234932363033295,
0.13632027804851532,
-0.05843184515833855,
-0.13171182572841644,
-0.02820662222802639,
0.09477535635232925,
0.04355064406991005,
-0.060202740132808685,
-0.1276080310344696,
-0.0682656466960907,
-0.03270931541919708,
-0.0022935839369893074,
-0.04131745919585228,
-0.002290143631398678,
-0.011816466227173805,
0.08523242175579071,
-0.08333887904882431,
-0.05337991565465927,
-0.020235886797308922,
-0.06682160496711731,
0.16135647892951965,
0.018354052677750587,
0.013679140247404575,
-0.013255320489406586,
0.07744071632623672,
-0.1265559196472168,
-0.1336318850517273,
-0.05428304150700569,
-0.03462158143520355,
-0.05692019686102867,
-0.05218517407774925,
0.01661340333521366,
-0.07750479131937027,
0.08107020705938339,
0.175857275724411,
-0.11741548776626587,
0.06220667064189911,
-0.009697522968053818,
0.030903007835149765,
0.024629943072795868,
0.08424429595470428,
-0.08733733743429184,
-0.08122188597917557,
0.04335138201713562,
0.05078619346022606,
0.05994807183742523,
0.0019238290842622519,
0.00010135882621398196,
-0.02575274370610714,
0.0019080113852396607,
0.07156599313020706,
0.06952427327632904,
0.029285242781043053,
-0.07679720222949982,
-0.05144278332591057,
0.17693321406841278,
-0.10877041518688202,
0.05875063315033913,
0.038095518946647644,
-0.010931058786809444,
-0.015534471720457077,
0.026578398421406746,
-0.028613878414034843,
-0.05422482639551163,
0.014574683271348476,
-0.052837468683719635,
0.006087095942348242,
-0.04121243208646774,
-0.04512706398963928,
0.052960656583309174,
0.044456131756305695,
-0.05848181992769241,
-0.12249575555324554,
-0.08548451960086823,
-0.056048691272735596,
0.033982738852500916,
-0.07165446132421494,
0.009947382844984531,
0.03666377440094948,
-0.08962082862854004,
0.029198521748185158,
0.02603435143828392,
0.022042686119675636,
-0.022196292877197266,
0.013299040496349335,
0.01655365526676178,
0.02530047670006752,
-0.031476203352212906,
-0.009267413057386875,
-0.05801313742995262,
0.08978775143623352,
-0.19187140464782715,
0.07319818437099457,
-0.06753759831190109,
-0.0458916574716568,
-0.1367833912372589,
-0.022929007187485695,
0.0001557070208946243,
-0.021685365587472916,
0.08853033930063248,
0.14028236269950867,
-0.15254990756511688,
-0.01800709031522274,
0.133183553814888,
-0.09939759224653244,
-0.08947071433067322,
0.09990517795085907,
0.010765177197754383,
-0.005259351339191198,
0.010271748527884483,
0.08585681021213531,
0.12536770105361938,
-0.056073881685733795,
-0.0745566338300705,
-0.10543492436408997,
-0.02353603020310402,
0.10377749055624008,
0.036487940698862076,
-0.06648851186037064,
0.006608173716813326,
0.0013413212727755308,
-0.03926616534590721,
-0.02470051497220993,
-0.01713505946099758,
-0.04516416788101196,
-0.051653604954481125,
-0.03216031193733215,
0.019755378365516663,
-0.0311356820166111,
-0.04306983947753906,
-0.06874691694974899,
-0.0662546455860138,
-0.03937284275889397,
0.10621660202741623,
0.0002958594122901559,
0.013465981930494308,
-0.06993132084608078,
0.12132570147514343,
-0.043282829225063324,
0.019509179517626762,
-0.13229045271873474,
-0.062326252460479736,
0.05427157133817673,
-0.17642687261104584,
0.057915039360523224,
-0.07802051305770874,
0.07007356733083725,
0.08792263269424438,
-0.03142979368567467,
-0.04443249851465225,
0.014312747865915298,
-0.011503690853714943,
-0.05503445118665695,
-0.13744132220745087,
-0.05339666083455086,
-0.03718583658337593,
0.1360798180103302,
-0.07717297226190567,
0.03823299705982208,
0.04849797114729881,
0.17298123240470886,
0.013948954641819,
-0.06230805069208145,
0.059276606887578964,
0.013416923582553864,
-0.01901276968419552,
-0.04254152253270149,
0.022224929183721542,
-0.015195275656878948,
-0.06335578858852386,
0.06797880679368973,
-0.16682805120944977,
-0.09226492792367935,
0.08205856382846832,
0.11543388664722443,
-0.10825490206480026,
-0.0011853284668177366,
-0.0016989412251859903,
-0.03144387900829315,
-0.02349192649126053,
-0.07283974438905716,
0.139337420463562,
0.03313908725976944,
0.05388158559799194,
-0.05538925528526306,
-0.07884406298398972,
0.012225146405398846,
-0.024691954255104065,
-0.07743193954229355,
0.12690365314483643,
0.02291349321603775,
-0.19036413729190826,
0.1086549460887909,
0.10671419650316238,
0.024346476420760155,
0.12972186505794525,
0.010163333266973495,
-0.020192109048366547,
-0.11323220282793045,
0.07214321941137314,
0.06328365206718445,
0.08764441311359406,
0.009260048158466816,
0.06931837648153305,
0.03807743266224861,
0.00207220995798707,
0.013925828970968723,
-0.11039725691080093,
0.03326145187020302,
0.019972439855337143,
-0.035299137234687805,
0.07873526960611343,
0.04182936251163483,
0.027880189940333366,
0.08705969154834747,
0.017179058864712715,
0.07423700392246246,
-0.01186748594045639,
-0.03507009521126747,
-0.09978814423084259,
0.13680647313594818,
-0.10421305149793625,
-0.1235826313495636,
-0.15689323842525482,
0.023495908826589584,
-0.1045532375574112,
-0.031237754970788956,
0.04314960539340973,
-0.04964744672179222,
-0.04190392047166824,
-0.09863945841789246,
0.07575149089097977,
0.04664899781346321,
-0.031183557584881783,
0.029061678797006607,
-0.04268911853432655,
0.05956493318080902,
-0.1079893708229065,
-0.03650837019085884,
0.0311584435403347,
0.003986986353993416,
0.03826166316866875,
0.06534901261329651,
0.09445244818925858,
0.06816873699426651,
0.010970757342875004,
-0.007111722603440285,
0.0249604694545269,
0.2451404184103012,
-0.09102492034435272,
0.08363116532564163,
0.1731153130531311,
0.012367261573672295,
0.08963974565267563,
0.1418784111738205,
0.0519220232963562,
-0.06685937941074371,
-0.013835576362907887,
0.023638667538762093,
-0.024145755916833878,
-0.19078323245048523,
-0.061981018632650375,
-0.049455784261226654,
0.02179974503815174,
0.016594912856817245,
0.056652046740055084,
-0.03246590495109558,
0.04957956075668335,
-0.06632956862449646,
-0.023145724087953568,
0.04272812604904175,
0.07885605841875076,
0.04942423850297928,
0.05232582241296768,
0.032079048454761505,
-0.030630147084593773,
-0.01874661073088646,
0.10800144821405411,
0.002024435205385089,
0.0828050896525383,
0.005763496737927198,
0.15540826320648193,
0.042188387364149094,
0.05763785168528557,
-0.006694222800433636,
0.05031910166144371,
0.012587900273501873,
0.01727849431335926,
0.011400274932384491,
-0.0922514945268631,
0.017045721411705017,
0.037224024534225464,
0.031022431328892708,
0.08605832606554031,
-0.019477766007184982,
-0.042653512209653854,
0.06828916072845459,
0.20636118948459625,
0.054764457046985626,
-0.18104644119739532,
-0.052608687430620193,
0.05323893949389458,
-0.005172972101718187,
-0.0441189780831337,
-0.0574423223733902,
0.0601525716483593,
-0.1421785205602646,
0.10297753661870956,
-0.02837306633591652,
0.08344317972660065,
-0.0768112912774086,
-0.025202374905347824,
-0.002551008015871048,
0.14314106106758118,
-0.007685830816626549,
0.009870217181742191,
-0.11473074555397034,
0.09677925705909729,
0.047673650085926056,
0.050363969057798386,
-0.0467650331556797,
0.05014784261584282,
0.06751372665166855,
-0.016121957451105118,
0.1009698212146759,
0.015580941922962666,
-0.08597638458013535,
-0.13121242821216583,
-0.07546716183423996,
-0.02328302152454853,
0.09724711626768112,
-0.07956860214471817,
0.11797893047332764,
-0.006677214056253433,
-0.041914381086826324,
-0.051285892724990845,
0.08619783818721771,
-0.16726025938987732,
-0.10657192021608353,
0.08049735426902771,
0.035933516919612885,
-0.00293326354585588,
-0.07889305055141449,
-0.03560107201337814,
-0.06792481988668442,
0.20780132710933685,
-0.12493043392896652,
-0.06458897888660431,
-0.12241725623607635,
-0.02887725457549095,
0.14838698506355286,
-0.09519194811582565,
0.016095273196697235,
-0.0620892271399498,
0.09675031155347824,
-0.05750473216176033,
-0.09322291612625122,
0.03633593022823334,
-0.0772782415151596,
-0.22163733839988708,
-0.029905345290899277,
0.1838245987892151,
0.031322285532951355,
0.03489852324128151,
0.0206821970641613,
0.04595774784684181,
0.021382097154855728,
-0.10996323078870773,
0.04105592891573906,
0.12817256152629852,
0.020900849252939224,
0.09039270132780075,
-0.045222893357276917,
-0.14004118740558624,
-0.09013182669878006,
-0.05677556246519089,
0.0997968390583992,
0.29861730337142944,
-0.05223260074853897,
0.030688846483826637,
0.04147891700267792,
-0.07043230533599854,
-0.15785279870033264,
-0.038915667682886124,
0.04657349735498428,
0.01753225363790989,
0.01858132891356945,
-0.07110119611024857,
0.0366257019340992,
0.10373353213071823,
-0.01873767003417015,
0.15667074918746948,
-0.263369619846344,
-0.12704867124557495,
-0.006682985462248325,
0.06700930744409561,
-0.01957916095852852,
-0.1452975571155548,
-0.11674544960260391,
-0.03796976059675217,
-0.1995268017053604,
0.09981389343738556,
-0.02467156946659088,
0.0965239554643631,
-0.023898916319012642,
0.015778150409460068,
0.03433018922805786,
-0.059150923043489456,
0.19815395772457123,
-0.039659541100263596,
0.04237253591418266,
-0.08535932749509811,
0.011777906678617,
0.026407601311802864,
-0.05703800171613693,
0.0950261577963829,
-0.035693515092134476,
0.017066331580281258,
-0.07961391657590866,
0.0030619590543210506,
-0.08434901386499405,
0.06988205015659332,
-0.053291499614715576,
-0.002292127348482609,
-0.06203543022274971,
0.07322024554014206,
0.02764309197664261,
-0.02415998838841915,
0.024202430620789528,
-0.021172653883695602,
0.10087262094020844,
0.2350863218307495,
0.09775305539369583,
0.05048970505595207,
-0.06141071021556854,
-0.00026774543221108615,
-0.04018312692642212,
0.02579716220498085,
-0.035933900624513626,
0.04462524875998497,
0.08787034451961517,
0.02455063909292221,
0.10782863199710846,
0.01810472458600998,
-0.1601009964942932,
0.018065448850393295,
0.06646080315113068,
-0.13265873491764069,
-0.2303391844034195,
0.003492377931252122,
0.04154038056731224,
-0.05837193876504898,
0.01806606538593769,
0.1862204670906067,
0.0009367137681692839,
-0.04963487386703491,
0.022825637832283974,
0.05416732281446457,
-0.03536570817232132,
0.15667866170406342,
0.00837104581296444,
0.04538920149207115,
-0.061773963272571564,
0.07580430060625076,
0.04042122885584831,
0.0016077554319053888,
0.018714869394898415,
0.17067234218120575,
-0.05190492421388626,
-0.07857772707939148,
0.04102415591478348,
0.03600002080202103,
-0.021241383627057076,
-0.021828800439834595,
-0.027304736897349358,
-0.10624261945486069,
0.013062734156847,
0.12654060125350952,
0.02681851200759411,
-0.018991822376847267,
0.08323623985052109,
-0.004140008706599474,
-0.028763674199581146,
0.07244459539651871,
0.05491362512111664,
0.08497325330972672,
-0.059819940477609634,
0.06284529715776443,
-0.02697847969830036,
0.03915136307477951,
-0.0026654857210814953,
0.019256316125392914,
-0.12592482566833496,
-0.05442740023136139,
-0.08961838483810425,
0.005896841175854206,
-0.11454344540834427,
-0.029031215235590935,
-0.0207052081823349,
-0.00936585757881403,
-0.02115669846534729,
0.03387326002120972,
-0.051367443054914474,
-0.11562348157167435,
-0.07470326870679855,
0.08342844247817993,
-0.14418748021125793,
-0.01641484536230564,
0.06919168680906296,
-0.07516539096832275,
0.09250383824110031,
0.01708120107650757,
0.0034909669775515795,
-0.009268661960959435,
-0.11126343160867691,
-0.02617836929857731,
-0.013749503530561924,
0.020847713574767113,
0.02727784588932991,
-0.20117603242397308,
-0.002813884522765875,
-0.04270019009709358,
-0.03368119150400162,
0.01241485308855772,
0.15425018966197968,
-0.09736622869968414,
-0.040856342762708664,
-0.03006616234779358,
-0.044825147837400436,
-0.06249745562672615,
0.018799860030412674,
0.10122360289096832,
0.009257921017706394,
0.08748460561037064,
-0.05421772226691246,
0.07078760117292404,
-0.17954121530056,
-0.01305952575057745,
0.022268623113632202,
-0.06352261453866959,
0.042893052101135254,
0.006506424397230148,
0.07432542741298676,
-0.024414660409092903,
0.033989157527685165,
-0.08212705701589584,
-0.018906600773334503,
0.026504697278141975,
-0.028520310297608376,
-0.021283207461237907,
0.017214683815836906,
0.020455483347177505,
-0.022489050403237343,
-0.024124853312969208,
-0.0313778817653656,
0.020932979881763458,
-0.019497908651828766,
-0.05568772181868553,
0.1307651400566101,
0.10116901993751526,
0.042394958436489105,
0.06033344939351082,
0.0699099600315094,
-0.06935380399227142,
-0.019917812198400497,
-0.015662092715501785,
-0.0501786470413208,
0.05921950191259384,
-0.0570748895406723,
0.11598431318998337,
0.0971740260720253,
-0.16128812730312347,
0.08817395567893982,
-0.03033575415611267,
-0.025268414989113808,
-0.0654015764594078,
-0.1747601330280304,
-0.06647072732448578,
-0.021373050287365913,
-0.007110625971108675,
-0.07304833084344864,
0.04277768358588219,
0.03683268278837204,
0.0017412106972187757,
-0.0061371480114758015,
0.12685923278331757,
-0.04393889009952545,
-0.058910902589559555,
0.035390980541706085,
0.037289973348379135,
-0.015837162733078003,
0.04472239688038826,
-0.006923268549144268,
0.022451741620898247,
0.012978330254554749,
0.05661128833889961,
0.08223818242549896,
-0.00805328506976366,
0.02584826387465,
-0.028271563351154327,
-0.10848145186901093,
-0.0006025367183610797,
0.005287060514092445,
0.01833254098892212,
0.08337931334972382,
0.05019523203372955,
0.016249405220150948,
-0.015350534580647945,
0.0810377299785614,
-0.02552863396704197,
-0.014691580086946487,
-0.1009829118847847,
0.15841826796531677,
-0.0559045635163784,
0.002254345454275608,
-0.0015080315060913563,
-0.11343378573656082,
-0.011597293429076672,
0.12920117378234863,
0.16337698698043823,
-0.08000685274600983,
0.03376281261444092,
0.017094949260354042,
0.00689410325139761,
0.004397694021463394,
0.049244899302721024,
0.059573933482170105,
0.10981784015893936,
-0.046494387090206146,
0.12055498361587524,
-0.0517595037817955,
-0.032632842659950256,
-0.07947799563407898,
0.09397511184215546,
-0.047008223831653595,
0.014640634879469872,
-0.017156826332211494,
0.07035721093416214,
-0.032913386821746826,
-0.16172361373901367,
0.03033449873328209,
-0.12312323600053787,
-0.138545423746109,
-0.00011208858632016927,
-0.0077992831356823444,
0.001197756384499371,
0.07809285819530487,
0.009989638812839985,
0.010767842642962933,
0.13262617588043213,
0.004294542595744133,
-0.07419104129076004,
-0.056784018874168396,
0.0005215644487179816,
-0.09722831845283508,
0.18984216451644897,
0.024254996329545975,
0.04178498685359955,
0.13091936707496643,
-0.033293675631284714,
-0.15407414734363556,
0.05585271492600441,
0.0700296014547348,
-0.09110260009765625,
0.022162333130836487,
0.14707012474536896,
-0.005258307326585054,
0.030615337193012238,
0.07387614995241165,
-0.053392838686704636,
0.019069518893957138,
0.0722138062119484,
0.019377747550606728,
-0.10239636152982712,
0.08977871388196945,
-0.10703950375318527,
0.13002605736255646,
0.18838851153850555,
-0.03489691764116287,
0.012865101918578148,
-0.03789073973894119,
0.04828639328479767,
0.038723692297935486,
0.09065895527601242,
-0.034608691930770874,
-0.19317959249019623,
0.06994378566741943,
-0.023887865245342255,
0.08475272357463837,
-0.1766246110200882,
-0.1239243745803833,
-0.029406068846583366,
-0.021863967180252075,
-0.013726417906582355,
0.0984341949224472,
0.09909792244434357,
0.0073255919851362705,
-0.023081263527274132,
-0.1434156447649002,
-0.02458539977669716,
0.10146303474903107,
-0.08631070703268051,
-0.04105154797434807
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "alexsherstinsky/Mistral-7B-v0.1-sharded"} | null | SudiptoPramanik/MistraWithRef_RL_RL_ExtractiveSummary | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:alexsherstinsky/Mistral-7B-v0.1-sharded",
"region:us"
] | 2024-02-06T07:23:38+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-alexsherstinsky/Mistral-7B-v0.1-sharded #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-alexsherstinsky/Mistral-7B-v0.1-sharded #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
45,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-alexsherstinsky/Mistral-7B-v0.1-sharded #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.110109843313694,
0.20426276326179504,
-0.0033242744393646717,
0.02683502435684204,
0.07316219061613083,
0.012933526188135147,
0.07240378856658936,
0.13029123842716217,
0.0330861434340477,
0.12623506784439087,
0.0663594976067543,
0.11143319308757782,
0.11400832235813141,
0.21561922132968903,
-0.003767105983570218,
-0.17360882461071014,
0.020239561796188354,
-0.05931909754872322,
0.034331753849983215,
0.12480589002370834,
0.13765163719654083,
-0.09187468141317368,
0.07441461831331253,
-0.023067325353622437,
-0.007257964462041855,
-0.030011918395757675,
-0.06413144618272781,
-0.012694036588072777,
0.052279502153396606,
0.0378146693110466,
0.060410551726818085,
-0.007752938661724329,
0.08189915865659714,
-0.2682739794254303,
0.016325049102306366,
0.04730505868792534,
-0.011336163617670536,
0.08267117291688919,
0.10176274180412292,
-0.04810560867190361,
0.1253533810377121,
-0.037304703146219254,
0.1296939104795456,
0.08034759759902954,
-0.10833422094583511,
-0.22102825343608856,
-0.06735025346279144,
0.0868915244936943,
0.1795286238193512,
0.06535330414772034,
-0.04263977333903313,
0.12261921912431717,
-0.055373914539813995,
0.028244679793715477,
0.08744969964027405,
-0.11619655787944794,
-0.067912258207798,
0.06995866447687149,
0.13108506798744202,
0.08480945229530334,
-0.11932659894227982,
-0.03515271842479706,
0.033694859594106674,
0.0460745133459568,
0.06793062388896942,
0.007016418967396021,
0.157415509223938,
0.02952895313501358,
-0.1433580219745636,
-0.04947032406926155,
0.10342932492494583,
0.007777987979352474,
-0.04197976365685463,
-0.21659140288829803,
-0.010902638547122478,
-0.09021607786417007,
-0.03760097175836563,
-0.04981773719191551,
0.0327930748462677,
0.01047485787421465,
0.12073933333158493,
-0.05324931815266609,
-0.08308704942464828,
-0.009831502102315426,
0.11944828927516937,
0.06926008313894272,
0.010240095667541027,
-0.019501550123095512,
0.00468682823702693,
0.11847840994596481,
0.05970883369445801,
-0.12965521216392517,
-0.062441613525152206,
-0.0627671331167221,
-0.029873453080654144,
-0.024195287376642227,
0.04842666909098625,
0.02648993954062462,
0.042996060103178024,
0.27919477224349976,
-0.02499905787408352,
0.06445452570915222,
0.03833015263080597,
0.017506811767816544,
0.0209523793309927,
0.10952203720808029,
-0.027887746691703796,
-0.18805280327796936,
-0.00785139575600624,
0.1037183478474617,
0.00951753742992878,
-0.03065507300198078,
-0.05571940541267395,
0.02089713327586651,
0.04004978388547897,
0.12380597740411758,
0.10219241678714752,
-0.028415348380804062,
-0.0656670331954956,
-0.056997787207365036,
0.20572279393672943,
-0.15231499075889587,
0.057325515896081924,
0.03066612035036087,
0.000978308031335473,
-0.06201085448265076,
0.011775513179600239,
0.008082118816673756,
-0.03692951425909996,
0.09554214775562286,
-0.06441154330968857,
-0.038193073123693466,
-0.11660514026880264,
-0.04551363363862038,
0.0367552787065506,
-0.020857995375990868,
-0.04893016442656517,
-0.03855219483375549,
-0.08102472871541977,
-0.10572165250778198,
0.10145532339811325,
-0.054616548120975494,
-0.04785343259572983,
-0.03009726293385029,
-0.07338356971740723,
0.02454744465649128,
0.029810938984155655,
0.061975590884685516,
-0.026527537032961845,
0.044389721006155014,
-0.02305789664387703,
0.07123309373855591,
0.07824490964412689,
0.033632248640060425,
-0.08152387291193008,
0.06508684158325195,
-0.17742902040481567,
0.07929858565330505,
-0.062077075242996216,
0.033570464700460434,
-0.16534066200256348,
0.0007017551106400788,
0.004242916591465473,
0.03086933307349682,
0.05304974690079689,
0.15916115045547485,
-0.2037542313337326,
-0.0345250740647316,
0.17959627509117126,
-0.10244475305080414,
-0.11975224316120148,
0.03716188296675682,
-0.04131937026977539,
0.17497402429580688,
0.04257720708847046,
0.02343027852475643,
0.09054774045944214,
-0.15291258692741394,
-0.01978994719684124,
-0.031267520040273666,
0.014165452681481838,
0.0622839592397213,
0.07420126348733902,
-0.08111037313938141,
0.00047640083357691765,
0.005789510440081358,
-0.05705531686544418,
-0.021126434206962585,
-0.038276441395282745,
-0.09720899909734726,
0.008260915987193584,
-0.07907192409038544,
0.007491591852158308,
0.0029457921627908945,
-0.0937480553984642,
-0.01217607595026493,
-0.14669319987297058,
-0.025165997445583344,
0.07373563945293427,
0.003158573294058442,
-0.005840174853801727,
-0.08364780247211456,
0.04531695321202278,
-0.04651717469096184,
-0.012500773184001446,
-0.15108191967010498,
-0.0011525335721671581,
0.02214769273996353,
-0.14006948471069336,
0.012162993662059307,
-0.14084939658641815,
0.07386209070682526,
0.013708565384149551,
-0.05641097202897072,
-0.03984736651182175,
0.014070483855903149,
-0.014163503423333168,
-0.07220236212015152,
-0.22516661882400513,
-0.03213506564497948,
-0.05659179016947746,
0.1322016566991806,
-0.22785338759422302,
0.045285001397132874,
0.0014656432904303074,
0.11130665242671967,
0.012530438601970673,
-0.06164274737238884,
0.021894264966249466,
-0.056239765137434006,
-0.026936599984765053,
-0.071257583796978,
-0.003288415027782321,
0.0019245932344347239,
-0.030422238633036613,
0.02249748632311821,
-0.1347021609544754,
-0.07268781960010529,
0.087431900203228,
0.08374550938606262,
-0.14605119824409485,
0.005244195461273193,
-0.035594403743743896,
-0.057235270738601685,
-0.06911290436983109,
-0.07025576382875443,
0.07089311629533768,
0.051931776106357574,
0.049817245453596115,
-0.08891286700963974,
-0.07021832466125488,
-0.0014828727580606937,
-0.018785672262310982,
-0.023595018312335014,
0.12765425443649292,
0.07284100353717804,
-0.10459782183170319,
0.09470570087432861,
0.07270287722349167,
0.02697981894016266,
0.0994887426495552,
-0.007307672407478094,
-0.09977997839450836,
-0.03138745203614235,
0.05709034204483032,
0.02002185396850109,
0.15080347657203674,
-0.05995139852166176,
0.04183671623468399,
0.044179778546094894,
-0.04822225123643875,
0.036573950201272964,
-0.0953039899468422,
0.011296747252345085,
0.006154231261461973,
-0.01610223948955536,
0.02957276813685894,
-0.022515933960676193,
0.0028687939047813416,
0.09212246537208557,
0.06580419838428497,
0.02490052953362465,
0.010820328257977962,
-0.03876588121056557,
-0.1413283348083496,
0.17779025435447693,
-0.08620518445968628,
-0.2214423269033432,
-0.15582458674907684,
0.03303509205579758,
0.060906071215867996,
-0.010964687913656235,
0.03860332816839218,
-0.04540291801095009,
-0.08722398430109024,
-0.08975794911384583,
0.02394494228065014,
0.044791966676712036,
-0.05868641659617424,
-0.06814705580472946,
0.0333222895860672,
0.024159414693713188,
-0.13040386140346527,
0.023749617859721184,
0.04909688979387283,
-0.00034987536491826177,
-0.005188827868551016,
0.03241325169801712,
0.08447565883398056,
0.2065049111843109,
-0.0026565860025584698,
-0.00025402664323337376,
0.05427759140729904,
0.28095370531082153,
-0.15285564959049225,
0.12382353097200394,
0.12293418496847153,
-0.06467143446207047,
0.08293068408966064,
0.19206713140010834,
0.03356461226940155,
-0.08545693010091782,
0.015078777447342873,
0.035098034888505936,
-0.03581055626273155,
-0.26805275678634644,
-0.04256880655884743,
-0.024064278230071068,
-0.06810324639081955,
0.09134866297245026,
0.08128290623426437,
0.09407679736614227,
0.03345697373151779,
-0.0752536952495575,
-0.0801953673362732,
0.04511275887489319,
0.12162093818187714,
-0.050455618649721146,
0.017943501472473145,
0.08430361747741699,
-0.04937650263309479,
0.007476978003978729,
0.0860724225640297,
-0.01453783456236124,
0.13549737632274628,
0.057939957827329636,
0.11976924538612366,
0.07539068162441254,
0.062220171093940735,
0.0041070543229579926,
0.04859903082251549,
-0.009308382868766785,
0.026934761554002762,
0.014352924190461636,
-0.09536643326282501,
0.02176397480070591,
0.11496482044458389,
-0.000005564658295043046,
0.026965824887156487,
0.0200294591486454,
-0.07863914221525192,
0.039087630808353424,
0.2055603265762329,
0.035098589956760406,
-0.2077338844537735,
-0.08235063403844833,
0.06039128825068474,
-0.07191184908151627,
-0.15041004121303558,
-0.013588255271315575,
0.011162610724568367,
-0.15474948287010193,
0.019464297220110893,
-0.04525265842676163,
0.11251527070999146,
-0.06617746502161026,
-0.04283647984266281,
0.09621220827102661,
0.04985488951206207,
-0.04508543387055397,
0.03984001651406288,
-0.18906085193157196,
0.10983707010746002,
0.03138548135757446,
0.07320403307676315,
-0.08476196974515915,
0.08112826943397522,
-0.0009192607249133289,
-0.014862818643450737,
0.15639187395572662,
-0.0022264497820287943,
-0.07685430347919464,
-0.08429781347513199,
-0.06939982622861862,
-0.01644659787416458,
0.08421112596988678,
-0.1347922384738922,
0.07905580848455429,
-0.023771880194544792,
-0.03505420312285423,
-0.006907613482326269,
-0.09923306107521057,
-0.1062455028295517,
-0.16373415291309357,
0.05397550389170647,
-0.08048416674137115,
0.016202464699745178,
-0.07462850958108902,
-0.05205504596233368,
0.04472297057509422,
0.16162064671516418,
-0.20440195500850677,
-0.11403585225343704,
-0.14111113548278809,
-0.10442207753658295,
0.15391606092453003,
-0.05450672283768654,
0.08898589760065079,
-0.012385696172714233,
0.15804588794708252,
-0.01586785353720188,
-0.023965999484062195,
0.080719493329525,
-0.09016864746809006,
-0.1869654506444931,
-0.05040700361132622,
0.19184014201164246,
0.13546326756477356,
0.027443328872323036,
-0.012780127115547657,
0.029926499351859093,
-0.054875750094652176,
-0.1049741879105568,
0.02712760493159294,
0.13121086359024048,
0.06301523000001907,
-0.011507576331496239,
-0.03961675614118576,
-0.10761462152004242,
-0.06617847084999084,
-0.037552911788225174,
-0.013129501603543758,
0.21204674243927002,
-0.07131116837263107,
0.15669883787631989,
0.13428831100463867,
-0.06544041633605957,
-0.20483754575252533,
0.03753198683261871,
0.035385578870773315,
0.018236977979540825,
0.026067109778523445,
-0.19457732141017914,
0.07309531420469284,
-0.02141459286212921,
-0.07382556796073914,
0.17703941464424133,
-0.19694438576698303,
-0.13029493391513824,
0.09852635860443115,
0.01708853431046009,
-0.19515210390090942,
-0.1503404974937439,
-0.11023768782615662,
-0.01704263687133789,
-0.12067081779241562,
0.06257613003253937,
0.017869386821985245,
0.015143671073019505,
0.010425310581922531,
0.014392692595720291,
0.04404516518115997,
-0.04461923986673355,
0.19380077719688416,
-0.033769670873880386,
0.004635848104953766,
-0.05519195273518562,
-0.10341878980398178,
0.006317309103906155,
-0.06523780524730682,
0.11731242388486862,
-0.025350118055939674,
0.02443709224462509,
-0.15162475407123566,
-0.0441247783601284,
-0.06597283482551575,
0.02474210597574711,
-0.09377969801425934,
-0.08283840119838715,
-0.04770177975296974,
0.07903211563825607,
0.09041422605514526,
-0.017946941778063774,
0.037181757390499115,
-0.09022089838981628,
0.09729012101888657,
0.20138823986053467,
0.17175358533859253,
0.05355418846011162,
-0.045174892991781235,
0.027436474338173866,
-0.03736938163638115,
0.04741382226347923,
-0.223489910364151,
0.0372220054268837,
0.061571162194013596,
0.03466495871543884,
0.08293042331933975,
-0.003242778591811657,
-0.1632089763879776,
-0.08290805667638779,
0.08469267934560776,
-0.06166926026344299,
-0.16236701607704163,
-0.02590998262166977,
0.035343270748853683,
-0.20532868802547455,
-0.04493626952171326,
0.04636850208044052,
-0.019659148529171944,
-0.04349243640899658,
0.02372937835752964,
0.08205997943878174,
-0.02056795358657837,
0.09458784013986588,
0.08412887156009674,
0.09041876345872879,
-0.09277179092168808,
0.057113803923130035,
0.08248565346002579,
-0.018163925036787987,
0.019434349611401558,
0.14658372104167938,
-0.04009850695729256,
-0.037152811884880066,
0.07934527099132538,
0.11620350182056427,
-0.009561249054968357,
-0.04248342663049698,
0.01543293334543705,
-0.052525997161865234,
0.07117807865142822,
0.13478432595729828,
0.02001088857650757,
-0.010263524018228054,
0.0705440491437912,
0.02907724305987358,
-0.09167683124542236,
0.12405750900506973,
0.056559495627880096,
0.02387828752398491,
-0.02098764106631279,
-0.02241300418972969,
-0.017281534150242805,
-0.007748601958155632,
-0.01214271318167448,
-0.0017991504864767194,
-0.09985364228487015,
-0.00010818249575095251,
-0.11642701923847198,
0.023298067972064018,
-0.07721348851919174,
0.0023173552472144365,
0.015358264558017254,
-0.043099671602249146,
-0.0026699115987867117,
-0.008397807367146015,
-0.0753541886806488,
-0.0537470318377018,
-0.03262227028608322,
0.07683586329221725,
-0.14318186044692993,
0.02770192362368107,
0.07104771584272385,
-0.10877057909965515,
0.062038786709308624,
-0.007841352373361588,
0.015183927491307259,
0.005674487445503473,
-0.15054509043693542,
0.057142749428749084,
-0.024486707523465157,
-0.01725645363330841,
0.006643626838922501,
-0.1708439141511917,
-0.006305950228124857,
-0.051809463649988174,
-0.0719214603304863,
0.009533489122986794,
-0.015077232383191586,
-0.12315729260444641,
0.12094684690237045,
0.0017362898215651512,
-0.06565473973751068,
-0.014981938526034355,
0.05908970907330513,
0.07032697647809982,
-0.01675652526319027,
0.09807620942592621,
-0.028157642111182213,
0.08416876941919327,
-0.18083736300468445,
-0.0069732884876430035,
-0.014492437243461609,
0.03272204473614693,
-0.027742622420191765,
-0.041203975677490234,
0.05140266567468643,
-0.012445840053260326,
0.14870640635490417,
-0.003039774252101779,
0.07148155570030212,
0.04709630087018013,
0.006209264509379864,
0.03288114070892334,
0.07110495865345001,
0.060228172689676285,
-0.023647742345929146,
-0.014423592947423458,
0.026967685669660568,
0.0021891314536333084,
-0.04349159449338913,
-0.11992581933736801,
0.06914249807596207,
0.18553002178668976,
0.07927486300468445,
0.03253671154379845,
-0.0010351829696446657,
-0.12652595341205597,
-0.08451662212610245,
0.0840892568230629,
-0.006650243885815144,
-0.03133326768875122,
-0.06817347556352615,
0.22585780918598175,
0.14429612457752228,
-0.19503988325595856,
0.07997902482748032,
-0.04266722500324249,
-0.03197807818651199,
-0.13879773020744324,
-0.16527847945690155,
-0.05710893124341965,
-0.029848087579011917,
-0.035192620009183884,
-0.06382724642753601,
0.060045745223760605,
0.03486083075404167,
-0.0010674390941858292,
-0.009919303469359875,
0.10035693645477295,
0.019814493134617805,
-0.03784974664449692,
0.05296546593308449,
0.06662454456090927,
0.047949232161045074,
-0.08730470389127731,
0.012993368320167065,
0.0028715517837554216,
0.0020882494281977415,
0.062306344509124756,
0.029509833082556725,
-0.0573907233774662,
0.02843424677848816,
-0.016508016735315323,
-0.12342271208763123,
0.04832269996404648,
-0.005592701956629753,
-0.013650083914399147,
0.14911004900932312,
0.03406934067606926,
0.0032252620439976454,
-0.009044532664120197,
0.23616735637187958,
-0.06634797155857086,
-0.07783475518226624,
-0.12452603876590729,
0.08175703883171082,
-0.05504899099469185,
0.03096017800271511,
0.012054885737597942,
-0.12354312837123871,
0.014887244440615177,
0.1677706241607666,
0.12264934927225113,
-0.0039043996948748827,
0.009955290704965591,
0.040559761226177216,
0.010925398208200932,
-0.018291443586349487,
0.017071884125471115,
0.042369335889816284,
0.20928987860679626,
-0.07440926879644394,
0.07626623660326004,
-0.010689882561564445,
-0.06825059652328491,
-0.024058185517787933,
0.12240655720233917,
-0.013858120888471603,
-0.010844264179468155,
-0.061129167675971985,
0.1378892958164215,
-0.06585115939378738,
-0.21978257596492767,
0.05639028176665306,
-0.09162507206201553,
-0.13244394958019257,
-0.0419410765171051,
0.012980849482119083,
-0.02495395764708519,
0.017732052132487297,
0.06565281748771667,
-0.05679652467370033,
0.162874236702919,
0.027744030579924583,
-0.05252060294151306,
-0.10697993636131287,
0.05384554713964462,
-0.1422901302576065,
0.28400719165802,
0.021953271701931953,
0.03439553827047348,
0.10937730967998505,
-0.02023676596581936,
-0.1391659528017044,
0.010313108563423157,
0.10612982511520386,
-0.056747034192085266,
0.057422272861003876,
0.15852361917495728,
-0.007096059154719114,
0.12099551409482956,
0.0602305643260479,
-0.0642608031630516,
0.0335778146982193,
-0.05697423592209816,
-0.05715372413396835,
-0.12165220826864243,
0.07334904372692108,
-0.0846068412065506,
0.1480477899312973,
0.1273326873779297,
-0.07294914871454239,
-0.007207845337688923,
-0.01663087122142315,
0.07814791053533554,
0.021788625046610832,
0.12389518320560455,
0.013396648690104485,
-0.18079085648059845,
0.04512875899672508,
0.0093348678201437,
0.11271371692419052,
-0.22138561308383942,
-0.05580883100628853,
0.0416252501308918,
-0.019063347950577736,
-0.09296101331710815,
0.12022978067398071,
0.04253165423870087,
0.016048051416873932,
-0.02939530834555626,
-0.09239242970943451,
0.020172851160168648,
0.15283069014549255,
-0.09778184443712234,
-0.01708502136170864
] |
null | null | null |
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain).
# Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_path = "PATH_TO_THIS_REPO"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
device_map="auto",
torch_dtype='auto'
).eval()
# Prompt content: "hi"
messages = [
{"role": "user", "content": "hi"}
]
input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt')
output_ids = model.generate(input_ids.to('cuda'))
response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
# Model response: "Hello! How can I assist you today?"
print(response)
``` | {"license": "other", "tags": ["autotrain", "text-generation"], "widget": [{"text": "I love AutoTrain because "}]} | text-generation | karthikrathod/llm_repo_v8_10e | [
"safetensors",
"autotrain",
"text-generation",
"conversational",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-06T07:24:08+00:00 | [] | [] | TAGS
#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us
|
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit AutoTrain.
# Usage
| [
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
"TAGS\n#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n",
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
37,
29,
3
] | [
"passage: TAGS\n#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.# Usage"
] | [
-0.02089853025972843,
0.03890561684966087,
-0.000762980489525944,
0.037646014243364334,
0.12435931712388992,
-0.03151287883520126,
0.23112058639526367,
0.04494147002696991,
-0.0575568825006485,
-0.09741601347923279,
0.18740901350975037,
0.17386218905448914,
-0.04334506019949913,
0.18782994151115417,
-0.03842408210039139,
-0.23926758766174316,
0.025883177295327187,
-0.0299287848174572,
0.14973880350589752,
0.12130317836999893,
0.15229710936546326,
-0.0829242467880249,
0.05421588197350502,
0.0457366518676281,
-0.19744595885276794,
0.02559680864214897,
0.07502555847167969,
-0.12002695351839066,
0.1892649233341217,
0.040962137281894684,
0.11825616657733917,
0.03324944153428078,
0.1392887830734253,
-0.1323491781949997,
0.01648798957467079,
0.004352208226919174,
-0.015311143361032009,
0.05287393927574158,
0.06082003563642502,
-0.034274082630872726,
0.09492087364196777,
0.19268183410167694,
0.12143059074878693,
0.05840236321091652,
-0.11065401881933212,
0.010359742678701878,
-0.02585293911397457,
0.015595678240060806,
0.12488947808742523,
0.121797576546669,
-0.02974177710711956,
0.2112775444984436,
-0.15929573774337769,
0.0785667672753334,
-0.11720649152994156,
-0.27605608105659485,
-0.007311069872230291,
0.2076014280319214,
0.06324941664934158,
-0.01046263799071312,
-0.13386328518390656,
0.06509426236152649,
0.1174032911658287,
-0.009732136502861977,
0.052042946219444275,
-0.01771010085940361,
-0.05808677524328232,
-0.008316196501255035,
-0.07604839652776718,
0.004176823887974024,
0.2025483250617981,
-0.06435471028089523,
-0.025879809632897377,
-0.1353462189435959,
-0.023601124063134193,
0.04423265904188156,
0.00368077983148396,
-0.10752057284116745,
-0.027382109314203262,
0.10084833204746246,
-0.02734971046447754,
-0.029397934675216675,
-0.1505003720521927,
-0.052210669964551926,
-0.08283388614654541,
0.030309928581118584,
0.0009279148071072996,
0.005750878248363733,
-0.10405394434928894,
0.10598764568567276,
-0.014304609969258308,
-0.09590446949005127,
0.050552137196063995,
-0.10984646528959274,
0.032756756991147995,
-0.11620049923658371,
-0.022093212231993675,
-0.08695599436759949,
0.015334513038396835,
0.21623161435127258,
0.16516101360321045,
-0.003946542274206877,
-0.08353158086538315,
0.03163360059261322,
0.032285887748003006,
0.09010306745767593,
0.07819008082151413,
-0.03263101354241371,
0.06596504896879196,
-0.04041123762726784,
-0.023562058806419373,
-0.026206638664007187,
-0.185186967253685,
0.04729154333472252,
0.006137077696621418,
0.06225769594311714,
-0.07368145138025284,
0.0758923590183258,
-0.02453492395579815,
0.05138348415493965,
0.03385981172323227,
-0.024239709600806236,
0.033983007073402405,
-0.03501613065600395,
0.015362166799604893,
-0.10241638869047165,
0.031124519184231758,
0.13060276210308075,
0.041950587183237076,
0.10722701251506805,
-0.0850663036108017,
-0.03558005392551422,
-0.10486439615488052,
-0.04084291309118271,
0.007949413731694221,
0.032330259680747986,
0.054881513118743896,
-0.20490533113479614,
-0.2844090461730957,
-0.034244854003190994,
0.052770666778087616,
-0.01975797861814499,
-0.07832197844982147,
-0.08976242691278458,
0.02668369561433792,
0.05969720333814621,
-0.03685269504785538,
0.04373543709516525,
-0.022354818880558014,
0.035809289664030075,
-0.0757109671831131,
-0.0067244102247059345,
-0.05800308659672737,
0.007987656630575657,
-0.1394086480140686,
-0.03892948850989342,
-0.01018267311155796,
0.01908150501549244,
-0.03469295799732208,
0.16121862828731537,
-0.010288888588547707,
0.05076303705573082,
-0.05012427642941475,
0.0520540215075016,
0.0038348138332366943,
0.15402163565158844,
-0.12805858254432678,
0.004590215627104044,
0.16217437386512756,
-0.10571835935115814,
-0.11733518540859222,
0.10878685116767883,
-0.11078933626413345,
0.2556385099887848,
0.1126617044210434,
0.14406165480613708,
0.0280612725764513,
-0.12442860752344131,
0.12669576704502106,
0.03417041152715683,
-0.09001672267913818,
-0.027209481224417686,
0.0015774862840771675,
-0.029457205906510353,
-0.21803908050060272,
0.024427056312561035,
0.13007183372974396,
0.07568662613630295,
-0.038225483149290085,
-0.08753399550914764,
-0.013979305513203144,
-0.05888194218277931,
0.05481130629777908,
0.00985832791775465,
0.11558723449707031,
-0.08033457398414612,
-0.03330337256193161,
0.02695239707827568,
0.04780461639165878,
0.07386761158704758,
-0.06066657975316048,
-0.07480321824550629,
-0.03438110277056694,
-0.00005651484752888791,
-0.004678141791373491,
-0.06730625778436661,
-0.0526479035615921,
-0.017854172736406326,
0.14683830738067627,
0.04623232036828995,
0.09310559928417206,
0.03057941049337387,
0.04193659499287605,
-0.01995823159813881,
0.009528989903628826,
0.16668112576007843,
0.04636063799262047,
-0.1251319795846939,
-0.09489064663648605,
0.1198563277721405,
-0.07429909706115723,
0.1495225876569748,
-0.2573336362838745,
0.02191506139934063,
-0.1137506514787674,
0.08119326084852219,
-0.015024850144982338,
0.06582725048065186,
-0.07824977487325668,
0.01642789877951145,
-0.08536693453788757,
0.0042993673123419285,
0.06477862596511841,
0.05614956095814705,
-0.026179833337664604,
0.14061102271080017,
-0.15953490138053894,
0.20964255928993225,
0.1161319687962532,
-0.10498357564210892,
-0.11012911051511765,
-0.10380077362060547,
0.004991353023797274,
-0.005274149589240551,
-0.11000026762485504,
-0.0012808284955099225,
0.11501315236091614,
-0.051325228065252304,
0.184207946062088,
-0.02479202300310135,
-0.027814652770757675,
-0.022695103660225868,
-0.08917387574911118,
-0.004993697162717581,
-0.013311133719980717,
0.0878831148147583,
-0.22586707770824432,
0.1341700702905655,
0.12997865676879883,
-0.011201041750609875,
0.1878158301115036,
0.02932732366025448,
0.028099095448851585,
0.004460213240236044,
-0.03533336520195007,
-0.010984709486365318,
0.02327060140669346,
-0.05687986686825752,
-0.01642347313463688,
0.013465014286339283,
0.010788206942379475,
0.028979692608118057,
-0.1271466314792633,
-0.04724383354187012,
0.014977987855672836,
0.056155066937208176,
0.016029085963964462,
0.05752420425415039,
-0.08498586714267731,
0.06746458262205124,
-0.025121653452515602,
-0.13671542704105377,
0.11770213395357132,
0.01172768697142601,
-0.12705263495445251,
0.17182578146457672,
-0.09404783695936203,
-0.196224644780159,
-0.17304284870624542,
-0.13585984706878662,
0.026043228805065155,
0.08839208632707596,
0.06914421916007996,
-0.06822904944419861,
-0.06807959824800491,
-0.004135052673518658,
-0.12654997408390045,
0.019381104037165642,
-0.03188987448811531,
-0.09604258090257645,
0.057193055748939514,
-0.009717279113829136,
-0.11798624694347382,
-0.05032327026128769,
0.00789867714047432,
-0.06308624148368835,
0.0605158731341362,
-0.03089403733611107,
0.054746001958847046,
0.1381448656320572,
-0.011948119848966599,
0.023544736206531525,
-0.0395624041557312,
0.17897886037826538,
-0.08672381937503815,
-0.0006116208387538791,
0.09763624519109726,
-0.048962898552417755,
0.028884489089250565,
0.2265005260705948,
0.03182725980877876,
-0.06495069712400436,
0.07192723453044891,
-0.035681869834661484,
-0.05174829810857773,
-0.19448144733905792,
-0.11049490422010422,
-0.010373943485319614,
-0.010003382340073586,
0.0674663707613945,
0.04859880357980728,
0.2720578908920288,
0.12234988063573837,
0.059470195323228836,
0.016185441985726357,
0.04209032282233238,
0.08999012410640717,
0.13016381859779358,
-0.04774774983525276,
0.17109765112400055,
-0.06409438699483871,
-0.16133272647857666,
0.044327691197395325,
-0.027926357463002205,
0.051227767020463943,
0.17565013468265533,
-0.03614453971385956,
0.047351136803627014,
0.11210278421640396,
0.12826228141784668,
0.1061127632856369,
0.07705885171890259,
-0.06504974514245987,
-0.010043035261332989,
0.00019683393475133926,
-0.05370469391345978,
0.14862267673015594,
-0.023733152076601982,
-0.06846705824136734,
-0.031645484268665314,
0.010693936608731747,
0.04905892163515091,
0.049152228981256485,
0.03127843141555786,
-0.2666167616844177,
0.03436502441763878,
0.046095263212919235,
-0.06547010689973831,
-0.11317573487758636,
0.09948568791151047,
-0.021655220538377762,
-0.18608878552913666,
0.017802411690354347,
-0.025920318439602852,
0.09116440266370773,
0.04311057925224304,
0.05799582228064537,
-0.09219425916671753,
-0.0708162784576416,
-0.05113530531525612,
0.15323954820632935,
-0.35677093267440796,
0.21487660706043243,
-0.014043435454368591,
0.0690545067191124,
-0.11276184022426605,
0.0014416693011298776,
0.07986348122358322,
0.16165494918823242,
0.11833548545837402,
-0.05488691106438637,
-0.16898946464061737,
-0.09826766699552536,
-0.08969532698392868,
-0.007673082873225212,
0.013347413390874863,
0.003650940954685211,
-0.005118653643876314,
-0.11486039310693741,
-0.0005021608667448163,
0.04620593041181564,
-0.010058995336294174,
-0.1808961033821106,
-0.15823762118816376,
-0.02242000214755535,
0.044828031212091446,
0.10119049996137619,
-0.033685166388750076,
-0.051781389862298965,
-0.06033768132328987,
0.15737107396125793,
0.04368119686841965,
0.012251429259777069,
-0.12371376901865005,
-0.05173582211136818,
-0.06613845378160477,
-0.022030174732208252,
0.07524938881397247,
0.009389028884470463,
0.12098590284585953,
-0.09848834574222565,
-0.05622165650129318,
0.10000088065862656,
-0.12879306077957153,
-0.044098254293203354,
-0.12273328751325607,
0.050619933754205704,
-0.026867562904953957,
-0.004624411929398775,
0.12226194888353348,
0.04077878221869469,
-0.07747189700603485,
-0.06510289013385773,
-0.02182580530643463,
-0.02168603427708149,
0.040108900517225266,
-0.11854132264852524,
-0.10533714294433594,
-0.144134521484375,
-0.03266002982854843,
-0.12010640650987625,
0.22031773626804352,
0.1510319709777832,
-0.0889979898929596,
0.16045299172401428,
0.21687199175357819,
-0.09459521621465683,
-0.28949886560440063,
-0.06218516454100609,
-0.05762689933180809,
0.0012655822793021798,
0.056375544518232346,
-0.09276837855577469,
0.08377362787723541,
-0.004379333462566137,
-0.0921919122338295,
-0.03929101675748825,
-0.10597379505634308,
-0.1628357619047165,
0.24811773002147675,
-0.00695221871137619,
0.216319277882576,
-0.06675629317760468,
-0.04963424429297447,
-0.11837507039308548,
0.03226492181420326,
0.05033990368247032,
-0.08250661194324493,
0.04896571487188339,
0.05970872566103935,
0.07762710750102997,
0.03615579381585121,
-0.04023800045251846,
0.0499248206615448,
-0.07690990716218948,
0.07372726500034332,
-0.17243541777133942,
-0.051966533064842224,
0.0291034784168005,
-0.02003716491162777,
0.11406885087490082,
-0.03866045922040939,
0.04375878721475601,
-0.05661903694272041,
-0.07238272577524185,
0.012632071040570736,
0.06424806267023087,
-0.0111227473244071,
-0.12185013294219971,
0.0070838648825883865,
-0.003560643410310149,
0.004385150969028473,
-0.06248250603675842,
0.016781898215413094,
-0.031206920742988586,
0.15563493967056274,
0.15905016660690308,
0.2279939204454422,
-0.06940897554159164,
0.057850778102874756,
-0.026937630027532578,
-0.12084269523620605,
0.07881549000740051,
-0.060470253229141235,
0.010923074558377266,
0.05394923686981201,
-0.05505755916237831,
0.16708660125732422,
0.053299445658922195,
-0.0007490343996323645,
-0.015869995579123497,
0.15427231788635254,
-0.17436520755290985,
0.028647977858781815,
-0.08862833678722382,
0.15710654854774475,
0.04452139511704445,
-0.029634831473231316,
0.10007839649915695,
-0.07933120429515839,
-0.029322272166609764,
0.006951325573027134,
0.017015496268868446,
-0.03554573282599449,
0.05849390849471092,
0.046525198966264725,
0.024086007848381996,
-0.06793931126594543,
0.026535160839557648,
0.07079220563173294,
0.0025835877750068903,
0.04738464578986168,
0.013694006018340588,
-0.09493011981248856,
-0.1037706807255745,
0.031061364337801933,
0.2576681077480316,
-0.1639707237482071,
-0.08702236413955688,
0.009577915072441101,
-0.10157066583633423,
-0.0026154285296797752,
0.07413817942142487,
0.06880449503660202,
0.03655710443854332,
-0.042900752276182175,
-0.013874638825654984,
-0.11066316813230515,
0.0910448282957077,
-0.015328219160437584,
0.0348287932574749,
-0.14798195660114288,
0.07496067136526108,
-0.03132447972893715,
-0.008997730910778046,
-0.08787791430950165,
-0.033700209110975266,
-0.12531232833862305,
0.030435124412178993,
-0.08465003967285156,
-0.04313739016652107,
-0.05273820459842682,
-0.010747137479484081,
0.0678463876247406,
-0.010134257376194,
-0.017098618671298027,
-0.024644924327731133,
-0.08711723238229752,
0.032871875911951065,
0.004344973247498274,
0.04483238607645035,
-0.04674182087182999,
-0.01993880234658718,
0.037311747670173645,
-0.000004001267825515242,
0.06050976738333702,
0.022565992549061775,
-0.007758983410894871,
0.03770044445991516,
-0.15966764092445374,
0.01916838437318802,
0.06271649152040482,
0.0006143683567643166,
0.016977902501821518,
-0.03355167806148529,
-0.0018841095734387636,
0.0999053344130516,
0.030659453943371773,
0.03639167547225952,
0.01731853187084198,
-0.0949004739522934,
0.037301186472177505,
0.10677090287208557,
-0.14946091175079346,
-0.022807510569691658,
-0.05471193790435791,
-0.011145985685288906,
-0.057102054357528687,
0.22019965946674347,
-0.11838836222887039,
0.04698079079389572,
-0.032419852912425995,
0.03750695660710335,
-0.0519956611096859,
-0.10454028844833374,
-0.10880608856678009,
-0.10406296700239182,
-0.036173172295093536,
-0.0017616144614294171,
0.2634603977203369,
0.14614185690879822,
-0.007627400569617748,
0.04732783883810043,
0.06023077666759491,
0.09986170381307602,
-0.0000392909932998009,
0.1907200664281845,
0.09213747829198837,
-0.004819431807845831,
-0.12899689376354218,
0.07417719066143036,
0.025308500975370407,
-0.10945913195610046,
0.0014507247833535075,
0.0060352059081196785,
-0.07921634614467621,
0.04549342021346092,
0.061475154012441635,
-0.049655646085739136,
-0.10908256471157074,
-0.1897570788860321,
-0.11767365038394928,
0.014547701925039291,
-0.1141902431845665,
0.006054932717233896,
0.18083947896957397,
-0.06133390590548515,
-0.022032413631677628,
-0.09275112301111221,
-0.0474187396466732,
-0.2181331366300583,
-0.15545961260795593,
-0.10639044642448425,
-0.08368334919214249,
0.04896046221256256,
-0.020269649103283882,
0.05286030098795891,
0.018245011568069458,
0.03993610292673111,
-0.06763483583927155,
0.08721300959587097,
-0.10831692814826965,
0.004784486256539822,
-0.009881925769150257,
-0.04393337666988373,
0.01711859367787838,
-0.19800134003162384,
-0.01726091466844082,
-0.14271385967731476,
-0.025886263698339462,
-0.02414889633655548,
-0.03923075646162033,
0.0015599187463521957,
-0.00659944349899888,
-0.022216126322746277,
-0.007123332936316729,
-0.010187787935137749,
0.03588121011853218,
0.030142245814204216,
0.06735268235206604,
0.01930520497262478,
0.021639658138155937,
0.03718075901269913,
0.2173466682434082,
-0.03672509640455246,
-0.18076519668102264,
-0.13255588710308075,
0.22741390764713287,
0.023755958303809166,
0.12003876268863678,
-0.07047237455844879,
-0.003944313619285822,
0.0649246871471405,
0.3151680529117584,
0.27447304129600525,
-0.04221269488334656,
0.012944314628839493,
-0.03759029880166054,
-0.008687055669724941,
-0.0077759926207363605,
0.17214618623256683,
0.0111585957929492,
0.18692266941070557,
-0.061342377215623856,
0.057751890271902084,
-0.007795935031026602,
-0.07976683229207993,
-0.05004684627056122,
0.1371750831604004,
-0.034483592957258224,
-0.013111086562275887,
-0.017309419810771942,
0.08474326133728027,
-0.06475097686052322,
0.1650533229112625,
-0.12438745051622391,
-0.03197024017572403,
-0.04968215525150299,
0.050263699144124985,
0.1181311383843422,
-0.009911769069731236,
0.03671935200691223,
-0.030859731137752533,
-0.025431539863348007,
0.018659215420484543,
-0.03971736878156662,
-0.08324228972196579,
-0.040832240134477615,
0.07943736016750336,
0.018289517611265182,
0.24940812587738037,
-0.016860337927937508,
0.06924241781234741,
0.07830806821584702,
-0.0007601219112984836,
-0.08936040103435516,
0.1169457733631134,
0.010533611290156841,
-0.053996723145246506,
0.1200164407491684,
-0.016792241483926773,
0.008844620548188686,
-0.001643515657633543,
-0.006236417684704065,
-0.18588665127754211,
0.14857490360736847,
-0.09602080285549164,
-0.0948827937245369,
-0.05673005431890488,
0.13433516025543213,
-0.02555198408663273,
0.16195133328437805,
0.05283422768115997,
-0.02981109544634819,
0.0056883953511714935,
-0.020765170454978943,
0.06717022508382797,
-0.002720105228945613,
-0.10159162431955338,
-0.03101331554353237,
-0.19819441437721252,
-0.01870795525610447,
0.10115032643079758,
-0.025165937840938568,
-0.23734821379184723,
-0.07709009200334549,
-0.06396035850048065,
-0.031772181391716,
-0.12610237300395966,
0.06999877095222473,
0.20647278428077698,
0.019630368798971176,
-0.009499672800302505,
-0.12196175009012222,
-0.011895264498889446,
0.02409667894244194,
-0.028847014531493187,
-0.10832608491182327
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text2text-generation | Kishan11/test | [
"transformers",
"safetensors",
"mt5",
"text2text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T07:24:47+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.051328662782907486,
0.16320262849330902,
-0.005184381268918514,
0.02251756191253662,
0.0970073789358139,
0.014154063537716866,
0.06843415647745132,
0.1107407882809639,
-0.01975325681269169,
0.11416343599557877,
0.03334542736411095,
0.09905029088258743,
0.11230985820293427,
0.15210844576358795,
-0.0026883550453931093,
-0.22916288673877716,
0.04717505723237991,
-0.126259446144104,
-0.0391545407474041,
0.11665356904268265,
0.14936181902885437,
-0.10113005340099335,
0.07722076028585434,
-0.030988991260528564,
-0.0076387980952858925,
-0.032814256846904755,
-0.0584266260266304,
-0.04825444892048836,
0.04778217524290085,
0.07062895596027374,
0.06431572884321213,
0.005459979176521301,
0.09387511014938354,
-0.2666816711425781,
0.019808361306786537,
0.07136410474777222,
-0.0030179128516465425,
0.07510632276535034,
0.060065507888793945,
-0.07584504038095474,
0.09495777636766434,
-0.05221208930015564,
0.1462833732366562,
0.08274403214454651,
-0.09120061993598938,
-0.18708732724189758,
-0.09102199226617813,
0.10069762915372849,
0.182896688580513,
0.048136260360479355,
-0.02258199453353882,
0.09849601238965988,
-0.0877108946442604,
0.012430681847035885,
0.05295773223042488,
-0.0661652609705925,
-0.05250505730509758,
0.06206765025854111,
0.0807318314909935,
0.07389048486948013,
-0.1229020431637764,
-0.02052718587219715,
0.007085064426064491,
0.008346792310476303,
0.08409557491540909,
0.02241266705095768,
0.15073934197425842,
0.03937535360455513,
-0.12801966071128845,
-0.04859030246734619,
0.10540658980607986,
0.04346438869833946,
-0.04617396742105484,
-0.2521538734436035,
-0.02899259887635708,
-0.02667904645204544,
-0.028535552322864532,
-0.03899479657411575,
0.04266434162855148,
-0.008417289704084396,
0.08040466904640198,
-0.008577418513596058,
-0.07472068816423416,
-0.03689853474497795,
0.06301835924386978,
0.05782817304134369,
0.024571044370532036,
-0.012929320335388184,
0.009019458666443825,
0.11621864885091782,
0.10347475856542587,
-0.12458769977092743,
-0.05281980708241463,
-0.06449289619922638,
-0.08029068261384964,
-0.04488494247198105,
0.03654220327734947,
0.03706445172429085,
0.04768291115760803,
0.24621281027793884,
0.01506307628005743,
0.05642811954021454,
0.03732956945896149,
0.009937233291566372,
0.06307817250490189,
0.11257993429899216,
-0.06105290725827217,
-0.10008567571640015,
-0.027280353009700775,
0.08969353139400482,
0.010363527573645115,
-0.03775324299931526,
-0.055724311619997025,
0.06201998516917229,
0.016127094626426697,
0.12256785482168198,
0.09371091425418854,
0.005687997210770845,
-0.06957072764635086,
-0.06580306589603424,
0.1956067681312561,
-0.16264818608760834,
0.04759228602051735,
0.03654341772198677,
-0.041625015437603,
-0.0036778231151401997,
0.011432791128754616,
0.022557828575372696,
-0.02259793132543564,
0.09028376638889313,
-0.054762836545705795,
-0.039149291813373566,
-0.10884293168783188,
-0.03559769690036774,
0.03501499071717262,
0.010106389410793781,
-0.03266524150967598,
-0.032688479870557785,
-0.0845562070608139,
-0.06954096257686615,
0.09501254558563232,
-0.07321515679359436,
-0.05012572929263115,
-0.017229469493031502,
-0.0749860405921936,
0.025271635502576828,
0.020357489585876465,
0.08093643933534622,
-0.021111220121383667,
0.0403398796916008,
-0.05398182198405266,
0.06070630997419357,
0.11319857090711594,
0.03403271734714508,
-0.05239800736308098,
0.061304960399866104,
-0.2453044205904007,
0.09998688101768494,
-0.070001982152462,
0.055851489305496216,
-0.15144193172454834,
-0.02402091957628727,
0.049301061779260635,
0.005924122873693705,
-0.010987705551087856,
0.14017367362976074,
-0.21632637083530426,
-0.026949219405651093,
0.16230235993862152,
-0.09370499104261398,
-0.07971389591693878,
0.05612146109342575,
-0.051730554550886154,
0.1073049008846283,
0.03953683376312256,
-0.027500709518790245,
0.06375151872634888,
-0.1381090134382248,
0.00040212401654571295,
-0.04698554798960686,
-0.020267371088266373,
0.15640391409397125,
0.07842771708965302,
-0.07008927315473557,
0.07522880285978317,
0.023509828373789787,
-0.024433886632323265,
-0.045395378023386,
-0.017890313640236855,
-0.10874514281749725,
0.011851978488266468,
-0.06341767311096191,
0.017641864717006683,
-0.024686299264431,
-0.09168647229671478,
-0.02869567647576332,
-0.17302192747592926,
-0.020798463374376297,
0.08556832373142242,
-0.008557834662497044,
-0.019592439755797386,
-0.11647101491689682,
0.01107026357203722,
0.03302319347858429,
0.004175432957708836,
-0.1334623545408249,
-0.052139077335596085,
0.02637198194861412,
-0.164566308259964,
0.03547604754567146,
-0.058501359075307846,
0.05004860460758209,
0.03211123123764992,
-0.03329542279243469,
-0.02857385016977787,
0.019824394956231117,
0.006224640179425478,
-0.014512771740555763,
-0.24753068387508392,
-0.029559502378106117,
-0.022510329261422157,
0.1669003665447235,
-0.2150150090456009,
0.037831008434295654,
0.071714386343956,
0.15179769694805145,
0.009358161129057407,
-0.03916828706860542,
0.007046530023217201,
-0.07627426087856293,
-0.030073314905166626,
-0.061686236411333084,
-0.007639498449862003,
-0.035809602588415146,
-0.05275571718811989,
0.0509311743080616,
-0.17023569345474243,
-0.03118530660867691,
0.10080438107252121,
0.06444820761680603,
-0.13638457655906677,
-0.017996715381741524,
-0.03701688349246979,
-0.04423511400818825,
-0.05668354406952858,
-0.05839645862579346,
0.10215305536985397,
0.0586850680410862,
0.0463530607521534,
-0.06511835008859634,
-0.07926809787750244,
0.0017478249501436949,
-0.01636536978185177,
-0.024005327373743057,
0.09562139213085175,
0.07809679955244064,
-0.128550186753273,
0.09052938222885132,
0.10546009242534637,
0.08905062079429626,
0.09842319041490555,
-0.021411335095763206,
-0.08413682132959366,
-0.05219675227999687,
0.025027591735124588,
0.01852312870323658,
0.1322891265153885,
-0.011911415494978428,
0.0515037402510643,
0.0410870797932148,
-0.012500960379838943,
0.0133194075897336,
-0.08853328227996826,
0.031934138387441635,
0.034452520310878754,
-0.0191578920930624,
0.03937562555074692,
-0.03862793743610382,
0.02063792385160923,
0.08873654156923294,
0.046638816595077515,
0.04356502741575241,
0.015139794908463955,
-0.04773930460214615,
-0.11410948634147644,
0.16559910774230957,
-0.12778021395206451,
-0.23285086452960968,
-0.14290852844715118,
0.0011466923169791698,
0.033804524689912796,
-0.01184175442904234,
0.0015852059004828334,
-0.0642307698726654,
-0.11860045790672302,
-0.09271383285522461,
0.013562624342739582,
0.046084269881248474,
-0.08686772733926773,
-0.058901336044073105,
0.060286350548267365,
0.04077553004026413,
-0.14541929960250854,
0.019320974126458168,
0.050119441002607346,
-0.09147187322378159,
-0.010142568498849869,
0.08355879038572311,
0.06965820491313934,
0.1800171583890915,
0.01213023066520691,
-0.019780350849032402,
0.03484240919351578,
0.2154977172613144,
-0.13454559445381165,
0.11504307389259338,
0.14155790209770203,
-0.08801047503948212,
0.08181151747703552,
0.19823485612869263,
0.04090484231710434,
-0.10179063677787781,
0.032779160887002945,
0.019162407144904137,
-0.030266085639595985,
-0.24904344975948334,
-0.07182002812623978,
-0.0015384424477815628,
-0.057178061455488205,
0.07619637995958328,
0.08923010528087616,
0.08985473215579987,
0.014106076210737228,
-0.09460747241973877,
-0.08114413172006607,
0.05550093203783035,
0.10361701995134354,
0.013524936512112617,
-0.010953888297080994,
0.08832461386919022,
-0.03412410989403725,
0.021281907334923744,
0.09124857932329178,
0.0012303158873692155,
0.17275752127170563,
0.05731576681137085,
0.18292830884456635,
0.07830806821584702,
0.07006581127643585,
0.010954656638205051,
0.01338646374642849,
0.021367572247982025,
0.02768668904900551,
-0.003376738866791129,
-0.08766022324562073,
-0.012305330485105515,
0.11825862526893616,
0.07214555889368057,
0.015057150274515152,
0.009626545943319798,
-0.03731287643313408,
0.0816538855433464,
0.17225277423858643,
-0.004195837303996086,
-0.18073737621307373,
-0.06175117567181587,
0.0808240994811058,
-0.0942181721329689,
-0.0978553295135498,
-0.02438163384795189,
0.030480818822979927,
-0.17188774049282074,
0.025788409635424614,
-0.01786126010119915,
0.11369330435991287,
-0.13831548392772675,
-0.020397847518324852,
0.06577063351869583,
0.07205205410718918,
0.001269566360861063,
0.05965572968125343,
-0.160665825009346,
0.10748846083879471,
0.014300641603767872,
0.0694030150771141,
-0.09756353497505188,
0.09956212341785431,
-0.003883731784299016,
-0.013259840197861195,
0.13451461493968964,
0.008388550952076912,
-0.0767815113067627,
-0.07929567247629166,
-0.09308768808841705,
-0.010677417740225792,
0.1281345635652542,
-0.1477975994348526,
0.08526764810085297,
-0.032966434955596924,
-0.045358359813690186,
0.0018689122516661882,
-0.10302852839231491,
-0.126983180642128,
-0.18611262738704681,
0.05528288707137108,
-0.13427682220935822,
0.03672889992594719,
-0.10563333332538605,
-0.03443324565887451,
-0.03114200569689274,
0.18988727033138275,
-0.22595620155334473,
-0.06757284700870514,
-0.15271799266338348,
-0.09939435124397278,
0.1442657858133316,
-0.051644183695316315,
0.08306998014450073,
-0.00545003917068243,
0.1809411197900772,
0.0204471405595541,
-0.024139465764164925,
0.0990704819560051,
-0.09589311480522156,
-0.1944846659898758,
-0.0805567279458046,
0.15795263648033142,
0.13470378518104553,
0.03485332801938057,
-0.0028120747301727533,
0.035782139748334885,
-0.017412638291716576,
-0.1234898641705513,
0.022648973390460014,
0.17883644998073578,
0.06719345599412918,
0.024359386414289474,
-0.026545848697423935,
-0.11176556348800659,
-0.06726131588220596,
-0.032448120415210724,
0.03110574185848236,
0.1833028942346573,
-0.0720948874950409,
0.18599410355091095,
0.14474965631961823,
-0.058687858283519745,
-0.19672097265720367,
0.012748281471431255,
0.03311282768845558,
0.004999228753149509,
0.0346582867205143,
-0.20138630270957947,
0.08562208712100983,
0.00026585807790979743,
-0.050549790263175964,
0.1327323019504547,
-0.1709088832139969,
-0.15038524568080902,
0.06999209523200989,
0.037376705557107925,
-0.19945120811462402,
-0.1203133687376976,
-0.09113546460866928,
-0.05348551273345947,
-0.18588998913764954,
0.10221870243549347,
0.028081931173801422,
0.008173973299562931,
0.032946377992630005,
0.027908386662602425,
0.015304015949368477,
-0.0412941575050354,
0.1914987862110138,
-0.02477763034403324,
0.029256334528326988,
-0.08408133685588837,
-0.0700240284204483,
0.04688924551010132,
-0.05717688798904419,
0.07894738763570786,
-0.025546837598085403,
0.012113095261156559,
-0.10591532289981842,
-0.04247612506151199,
-0.031593095511198044,
0.014440414495766163,
-0.09696424007415771,
-0.08737108111381531,
-0.04703705385327339,
0.09430442005395889,
0.0985439270734787,
-0.03574737161397934,
-0.03532914072275162,
-0.07118727266788483,
0.04086564853787422,
0.19116361439228058,
0.17946277558803558,
0.04091198369860649,
-0.07925742864608765,
-0.006378510035574436,
-0.011479404754936695,
0.04257804900407791,
-0.2157202661037445,
0.06462159752845764,
0.050062380731105804,
0.019104523584246635,
0.1179836317896843,
-0.019891055300831795,
-0.15467524528503418,
-0.06981337070465088,
0.06141482666134834,
-0.05946635082364082,
-0.19295614957809448,
0.0024802349507808685,
0.055617205798625946,
-0.1677827537059784,
-0.04761936143040657,
0.0435972660779953,
-0.0037963632494211197,
-0.03886210918426514,
0.018091563135385513,
0.08966289460659027,
0.0023356645833700895,
0.07194776087999344,
0.057763248682022095,
0.08416751027107239,
-0.10276257991790771,
0.0799563005566597,
0.08647435903549194,
-0.08149342983961105,
0.02556447684764862,
0.09818926453590393,
-0.05909668654203415,
-0.031693510711193085,
0.027908844873309135,
0.08353206515312195,
0.015399634838104248,
-0.0418633408844471,
0.011942686513066292,
-0.10034026950597763,
0.06462868303060532,
0.09035695344209671,
0.030880169942975044,
0.013196618296205997,
0.03233819827437401,
0.046119559556245804,
-0.07282152771949768,
0.12195046246051788,
0.029250819236040115,
0.014950193464756012,
-0.0410873182117939,
-0.04673595353960991,
0.02354983240365982,
-0.02780505269765854,
-0.006134995259344578,
-0.0344335101544857,
-0.07343567907810211,
-0.017396869137883186,
-0.1646784543991089,
-0.014960045926272869,
-0.050430599600076675,
0.009284928441047668,
0.027561979368329048,
-0.03680562227964401,
0.0060029830783605576,
0.00910363718867302,
-0.07602906227111816,
-0.06694655120372772,
-0.0237664096057415,
0.09428686648607254,
-0.16284170746803284,
0.022817213088274002,
0.08406320959329605,
-0.11997158825397491,
0.09150046110153198,
0.019169924780726433,
-0.0043096113950014114,
0.025971177965402603,
-0.1505080759525299,
0.03534393385052681,
-0.03375111520290375,
0.013465344905853271,
0.045393940061330795,
-0.22764363884925842,
-0.0006596326129510999,
-0.03479163348674774,
-0.06353507936000824,
-0.009049107320606709,
-0.0390302799642086,
-0.11447589844465256,
0.10432159900665283,
0.00737398024648428,
-0.08732112497091293,
-0.03339998051524162,
0.03389183431863785,
0.08260936290025711,
-0.02571156620979309,
0.15404881536960602,
-0.00210543698631227,
0.0752558782696724,
-0.17060600221157074,
-0.01917494460940361,
-0.008899319916963577,
0.02190987393260002,
-0.01933358423411846,
-0.008870418183505535,
0.04324660077691078,
-0.02540694922208786,
0.18302802741527557,
-0.026055961847305298,
0.023916294798254967,
0.06704080104827881,
0.028971081599593163,
-0.026980608701705933,
0.10390723496675491,
0.05128125473856926,
0.01839378848671913,
0.019961433485150337,
0.007502324413508177,
-0.04053238779306412,
-0.023456495255231857,
-0.1998721808195114,
0.07054254412651062,
0.14529378712177277,
0.09391805529594421,
-0.01637977734208107,
0.08406981825828552,
-0.09878277778625488,
-0.11636010557413101,
0.1169997826218605,
-0.05138581991195679,
-0.002931964583694935,
-0.06808105856180191,
0.12546302378177643,
0.1463906168937683,
-0.19197741150856018,
0.07215752452611923,
-0.06787984073162079,
-0.0486978255212307,
-0.11575321853160858,
-0.19757696986198425,
-0.057408250868320465,
-0.051381710916757584,
-0.015996510162949562,
-0.047199856489896774,
0.07636301219463348,
0.05376458168029785,
0.008772914297878742,
-0.0009363067802041769,
0.06455569714307785,
-0.028812330216169357,
-0.00030995410634204745,
0.02975199744105339,
0.06359979510307312,
0.011518939398229122,
-0.029690392315387726,
0.019902538508176804,
-0.0106467604637146,
0.04097694158554077,
0.0645679235458374,
0.0467936247587204,
-0.030980709940195084,
0.01525089144706726,
-0.036990053951740265,
-0.10704203695058823,
0.041513945907354355,
-0.028119675815105438,
-0.07895249873399734,
0.1515752673149109,
0.023193838074803352,
0.007386222947388887,
-0.020481420680880547,
0.24136120080947876,
-0.07422091066837311,
-0.09643889963626862,
-0.14721740782260895,
0.10216663032770157,
-0.04222400486469269,
0.060273054987192154,
0.04530956223607063,
-0.10375474393367767,
0.01553369965404272,
0.1282518357038498,
0.16521716117858887,
-0.04190836846828461,
0.020468562841415405,
0.028186822310090065,
0.004011107608675957,
-0.036687396466732025,
0.04996533691883087,
0.06798537820577621,
0.15550518035888672,
-0.04796762019395828,
0.09592252224683762,
-0.00007890416600275785,
-0.0971902385354042,
-0.03596784546971321,
0.11750209331512451,
-0.017748000100255013,
0.01799829490482807,
-0.05343281477689743,
0.1203235387802124,
-0.06190110743045807,
-0.2302190214395523,
0.056667957454919815,
-0.06700639426708221,
-0.13796204328536987,
-0.02338138222694397,
0.07993879169225693,
-0.012393412180244923,
0.027088945731520653,
0.07425012439489365,
-0.07398337125778198,
0.19792741537094116,
0.03790948912501335,
-0.056375857442617416,
-0.05510881170630455,
0.08025650680065155,
-0.10229071229696274,
0.27599942684173584,
0.01627948135137558,
0.046000488102436066,
0.10458707809448242,
-0.01374407671391964,
-0.13688722252845764,
0.020522065460681915,
0.09743198752403259,
-0.09572482854127884,
0.04322853684425354,
0.2035953551530838,
-0.0014828984858468175,
0.12146259099245071,
0.0774778351187706,
-0.07723316550254822,
0.04730105772614479,
-0.09168743342161179,
-0.06985042244195938,
-0.09130857139825821,
0.09622733294963837,
-0.07444840669631958,
0.1427851915359497,
0.13178566098213196,
-0.05372144281864166,
0.0102944141253829,
-0.030041957274079323,
0.04689300060272217,
0.003170362673699856,
0.10121745616197586,
0.008502482436597347,
-0.18413890898227692,
0.022574106231331825,
0.013636616058647633,
0.10810644179582596,
-0.16641046106815338,
-0.09936638176441193,
0.044252779334783554,
-0.0010333930840715766,
-0.060718707740306854,
0.12968340516090393,
0.061216048896312714,
0.04452041536569595,
-0.041992466896772385,
-0.026066385209560394,
-0.00883333757519722,
0.13815481960773468,
-0.10480473190546036,
0.0015699166106060147
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.1 | {"library_name": "peft", "base_model": "openai/whisper-small"} | null | unanam/medi_lora_test | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:openai/whisper-small",
"region:us"
] | 2024-02-06T07:25:08+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-small #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.1 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-small #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
37,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-small #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.1"
] | [
-0.1076493114233017,
0.1983276903629303,
-0.0030863555148243904,
0.0321795716881752,
0.0915658101439476,
0.015891052782535553,
0.05633755400776863,
0.12364462018013,
-0.023953059688210487,
0.10961870849132538,
0.06933917105197906,
0.09555713087320328,
0.10782624781131744,
0.22033579647541046,
0.004158330149948597,
-0.20340892672538757,
0.030735481530427933,
-0.09517928212881088,
-0.0060378434136509895,
0.12312088161706924,
0.14362573623657227,
-0.09643791615962982,
0.07792773842811584,
-0.013907761313021183,
-0.0032701713498681784,
-0.03422047570347786,
-0.07384517788887024,
-0.03280039504170418,
0.04283977672457695,
0.04800018295645714,
0.054694920778274536,
-0.002007583389058709,
0.08344709128141403,
-0.26758667826652527,
0.019367258995771408,
0.04622700810432434,
-0.0050204754807055,
0.08665702491998672,
0.09662847965955734,
-0.043118398636579514,
0.12667696177959442,
-0.028781600296497345,
0.14164890348911285,
0.08002130687236786,
-0.0919370949268341,
-0.2328198254108429,
-0.0656222477555275,
0.08521860837936401,
0.17843151092529297,
0.07512778788805008,
-0.04341765120625496,
0.1328190267086029,
-0.08986219763755798,
0.018186595290899277,
0.043652087450027466,
-0.09434240311384201,
-0.06971664726734161,
0.051445215940475464,
0.10071490705013275,
0.054244764149188995,
-0.13309679925441742,
-0.030154060572385788,
0.026816044002771378,
0.03318323567509651,
0.0763542503118515,
0.01438731886446476,
0.14895938336849213,
0.023201987147331238,
-0.14727844297885895,
-0.04265831783413887,
0.13851937651634216,
0.026314957067370415,
-0.037391602993011475,
-0.22940459847450256,
0.0011385619873180985,
-0.07937060296535492,
-0.027850504964590073,
-0.04643959924578667,
0.03835776075720787,
0.0035814461298286915,
0.10350736975669861,
-0.025973044335842133,
-0.08995330333709717,
-0.013638122007250786,
0.09043186157941818,
0.05712338164448738,
0.024961039423942566,
-0.018793685361742973,
0.007855760864913464,
0.12285175174474716,
0.05301816761493683,
-0.13211382925510406,
-0.062122903764247894,
-0.07229094952344894,
-0.0466148816049099,
-0.04049689322710037,
0.040143758058547974,
0.0389413945376873,
0.06058909744024277,
0.2540028989315033,
-0.040138933807611465,
0.05887095257639885,
0.06300431489944458,
0.021149685606360435,
0.04726818948984146,
0.09386305510997772,
-0.057538650929927826,
-0.15375259518623352,
-0.014503112994134426,
0.09714944660663605,
-0.0015468598576262593,
-0.025122933089733124,
-0.0468185730278492,
0.045701492577791214,
0.032771091908216476,
0.1081102266907692,
0.09657766669988632,
-0.006575165782123804,
-0.07525230199098587,
-0.05242461338639259,
0.20518597960472107,
-0.1517454832792282,
0.040219634771347046,
0.021062158048152924,
-0.014131353236734867,
-0.047233957797288895,
0.012608767487108707,
0.01792128011584282,
-0.02555043436586857,
0.09977924823760986,
-0.06769528239965439,
-0.04050179198384285,
-0.11594671756029129,
-0.024581987410783768,
0.03249194473028183,
0.009308249689638615,
-0.02965298667550087,
-0.031701747328042984,
-0.06652764230966568,
-0.09290814399719238,
0.10246965289115906,
-0.06180013343691826,
-0.05941527336835861,
-0.029276642948389053,
-0.09009333699941635,
0.02061585895717144,
0.027388054877519608,
0.09028425067663193,
-0.026221640408039093,
0.04096124693751335,
-0.014384268783032894,
0.06348905712366104,
0.081583172082901,
0.032804589718580246,
-0.07547352463006973,
0.06447158753871918,
-0.19954754412174225,
0.0857921838760376,
-0.084710031747818,
0.03198399022221565,
-0.15881584584712982,
-0.018154140561819077,
0.0032789516262710094,
0.022762620821595192,
0.031502269208431244,
0.16389262676239014,
-0.2018553465604782,
-0.03459901735186577,
0.16319474577903748,
-0.11037147045135498,
-0.11824139207601547,
0.04280855134129524,
-0.046024423092603683,
0.15387937426567078,
0.02193201147019863,
-0.0015563314082100987,
0.09718871116638184,
-0.1482716202735901,
-0.024007121101021767,
-0.017556210979819298,
-0.0035372173879295588,
0.10268247872591019,
0.08908611536026001,
-0.08555474132299423,
0.02469658851623535,
0.014980979263782501,
-0.04626452922821045,
-0.026913143694400787,
-0.048375826328992844,
-0.10793200880289078,
0.00752988550812006,
-0.08077334612607956,
0.02308584563434124,
-0.005928411614149809,
-0.08180981129407883,
-0.012791378423571587,
-0.16180141270160675,
-0.037991784512996674,
0.08456020802259445,
0.008789405226707458,
-0.02098114602267742,
-0.10236939042806625,
0.04385245591402054,
-0.02773316204547882,
-0.019702883437275887,
-0.14682608842849731,
-0.022154204547405243,
0.01778225786983967,
-0.13768625259399414,
0.005281948484480381,
-0.117502860724926,
0.06859368085861206,
0.013534052297472954,
-0.0617329403758049,
-0.03721749410033226,
-0.00421797065064311,
0.005688815843313932,
-0.049892447888851166,
-0.244455024600029,
-0.021481839939951897,
-0.054059360176324844,
0.15850962698459625,
-0.22450770437717438,
0.038341864943504333,
0.050851497799158096,
0.12934011220932007,
0.0053882896900177,
-0.06182710453867912,
0.030630284920334816,
-0.06666679680347443,
-0.024592405185103416,
-0.07306734472513199,
-0.005906838923692703,
-0.005698047112673521,
-0.04731233790516853,
0.01670444756746292,
-0.1243169903755188,
-0.03097485937178135,
0.09924785792827606,
0.07058295607566833,
-0.1607288420200348,
-0.021585823968052864,
-0.04421459883451462,
-0.06136966496706009,
-0.08329130709171295,
-0.05805009976029396,
0.11161011457443237,
0.05035965517163277,
0.03847753256559372,
-0.07635572552680969,
-0.07006720453500748,
0.008144088089466095,
-0.02595604583621025,
-0.021186092868447304,
0.11548041552305222,
0.07015448808670044,
-0.11504032462835312,
0.0973467230796814,
0.07073502987623215,
0.03342175483703613,
0.0851202979683876,
-0.0280435960739851,
-0.10301949083805084,
-0.027422258630394936,
0.05016901716589928,
0.015683740377426147,
0.15817846357822418,
-0.06941792368888855,
0.05380067974328995,
0.04556833580136299,
-0.039361756294965744,
0.04572197049856186,
-0.0958394780755043,
0.008472136221826077,
0.00920481700450182,
-0.012628125958144665,
0.018263377249240875,
-0.019145743921399117,
0.009420912712812424,
0.08591506630182266,
0.049637455493211746,
0.03773082420229912,
0.02436952106654644,
-0.03237615153193474,
-0.13039547204971313,
0.1881663054227829,
-0.09735795855522156,
-0.23805677890777588,
-0.1566658616065979,
0.0619957372546196,
0.05775758624076843,
-0.01721619814634323,
0.023118281736969948,
-0.058119822293519974,
-0.10634229332208633,
-0.08314651250839233,
-0.0010542846284806728,
0.03128631040453911,
-0.05716496333479881,
-0.06432021409273148,
0.045638080686330795,
0.04694904386997223,
-0.11722119897603989,
0.03548428788781166,
0.05590042844414711,
-0.020242607221007347,
0.0030482348520308733,
0.05601544678211212,
0.08422736823558807,
0.182200089097023,
-0.004974318668246269,
-0.002466564066708088,
0.04954039677977562,
0.2783425450325012,
-0.16014724969863892,
0.10980077087879181,
0.1245369166135788,
-0.0698106661438942,
0.08021605759859085,
0.19261182844638824,
0.03494355082511902,
-0.10229311883449554,
0.028608860448002815,
0.027718408033251762,
-0.030323751270771027,
-0.26372426748275757,
-0.05259042978286743,
-0.015047707594931126,
-0.08474447578191757,
0.07769416272640228,
0.09034442901611328,
0.0787915512919426,
0.03926971182227135,
-0.07246412336826324,
-0.08992543816566467,
0.03577519580721855,
0.09965889155864716,
-0.01970483362674713,
0.005910307634621859,
0.08410564810037613,
-0.03590540215373039,
0.009793003089725971,
0.0971779152750969,
-0.015544719062745571,
0.1641339659690857,
0.04993084445595741,
0.10748257488012314,
0.0819782167673111,
0.08726797252893448,
-0.0005193806136958301,
0.02727576717734337,
0.011590331792831421,
0.022533316165208817,
0.01381439995020628,
-0.08577758818864822,
0.02547571063041687,
0.11214505881071091,
0.036740440875291824,
0.032772839069366455,
0.01444737333804369,
-0.03526489436626434,
0.0516301691532135,
0.17740067839622498,
0.009861274622380733,
-0.20111730694770813,
-0.08016648888587952,
0.06206071749329567,
-0.0795067697763443,
-0.1346665471792221,
-0.013888124376535416,
0.03803299739956856,
-0.16843293607234955,
0.027718214318156242,
-0.03985047712922096,
0.10041912645101547,
-0.0862351506948471,
-0.03858783841133118,
0.09548331797122955,
0.06746668368577957,
-0.026204491034150124,
0.05603724345564842,
-0.18718457221984863,
0.13196751475334167,
0.025866081938147545,
0.07103614509105682,
-0.0860164687037468,
0.10029779374599457,
0.004205841105431318,
0.0009366541053168476,
0.169908344745636,
0.002596179721876979,
-0.06030109524726868,
-0.0654691532254219,
-0.09628865867853165,
-0.012736977078020573,
0.09725938737392426,
-0.13308683037757874,
0.0665472075343132,
-0.01883644051849842,
-0.02848105877637863,
-0.0038482972886413336,
-0.08089105784893036,
-0.12055569887161255,
-0.17128720879554749,
0.055584464222192764,
-0.09696833789348602,
0.026417383924126625,
-0.0953107625246048,
-0.062148261815309525,
0.0023791391868144274,
0.1735740303993225,
-0.20987804234027863,
-0.10082574188709259,
-0.1499793976545334,
-0.09069482237100601,
0.16004957258701324,
-0.04506171494722366,
0.0834546834230423,
0.0002278044557897374,
0.16277861595153809,
0.012040389701724052,
-0.009224067442119122,
0.10199686884880066,
-0.09140172600746155,
-0.19608080387115479,
-0.056755732744932175,
0.16762030124664307,
0.13349591195583344,
0.037220459431409836,
-0.014349030330777168,
0.02760496363043785,
-0.0480252280831337,
-0.12105893343687057,
0.024897679686546326,
0.1442984640598297,
0.06282637268304825,
-0.011814827099442482,
-0.025748824700713158,
-0.10190238803625107,
-0.059641607105731964,
-0.045497749000787735,
0.0008954288205131888,
0.19032233953475952,
-0.07545129209756851,
0.16307900846004486,
0.11689619719982147,
-0.05564514175057411,
-0.20979732275009155,
0.045433931052684784,
0.054933637380599976,
0.01316364761441946,
0.03858482837677002,
-0.19467966258525848,
0.08807045966386795,
-0.0017842025263234973,
-0.07422445714473724,
0.16145873069763184,
-0.17291738092899323,
-0.1406075805425644,
0.10080746561288834,
0.03399902582168579,
-0.21785761415958405,
-0.1403006613254547,
-0.10279868543148041,
-0.018881503492593765,
-0.11589798331260681,
0.050210271030664444,
-0.002760865492746234,
0.012013448402285576,
0.02727854996919632,
0.01050734892487526,
0.026881346479058266,
-0.04779256507754326,
0.20390857756137848,
-0.031592391431331635,
0.011317810975015163,
-0.05312700569629669,
-0.08256140351295471,
0.023992329835891724,
-0.04610610753297806,
0.10245703160762787,
-0.006674810312688351,
0.026216629892587662,
-0.1524333357810974,
-0.042280346155166626,
-0.052563779056072235,
0.032902706414461136,
-0.09247459471225739,
-0.0882364884018898,
-0.043957602232694626,
0.0924544483423233,
0.0955497995018959,
-0.026336953043937683,
0.0027183894999325275,
-0.09096531569957733,
0.07398661226034164,
0.2018895149230957,
0.1966233253479004,
0.06757651269435883,
-0.0657537505030632,
0.021399857476353645,
-0.030358031392097473,
0.04847821593284607,
-0.23130162060260773,
0.04004419595003128,
0.05652376636862755,
0.021406229585409164,
0.08490227162837982,
-0.011100927367806435,
-0.15559959411621094,
-0.07046368718147278,
0.08477463573217392,
-0.051277246326208115,
-0.1720779985189438,
-0.026329487562179565,
0.03986504301428795,
-0.20511743426322937,
-0.038622498512268066,
0.021532917395234108,
-0.022605692967772484,
-0.03549124300479889,
0.022901972755789757,
0.07980462908744812,
-0.017224472016096115,
0.10847709327936172,
0.08394552022218704,
0.09379895776510239,
-0.10465157777070999,
0.07601074129343033,
0.07335831224918365,
-0.04453139007091522,
0.029397794976830482,
0.1152505949139595,
-0.0484611950814724,
-0.033827319741249084,
0.07841593772172928,
0.09077206999063492,
0.03283395990729332,
-0.054976046085357666,
0.013219248503446579,
-0.060914263129234314,
0.06021353602409363,
0.11719082295894623,
0.029192456975579262,
-0.004156169947236776,
0.05887904018163681,
0.03329429030418396,
-0.09060423821210861,
0.11029524356126785,
0.057404227554798126,
0.015794776380062103,
-0.046272993087768555,
-0.03435831889510155,
-0.0062603335827589035,
-0.016675807535648346,
-0.02058086358010769,
-0.007289220578968525,
-0.09135536104440689,
-0.0068361395969986916,
-0.09154845774173737,
0.025951888412237167,
-0.07359951734542847,
0.010295849293470383,
0.0286477692425251,
-0.05161719396710396,
0.001144910347647965,
0.0067473822273314,
-0.07389476150274277,
-0.04638208448886871,
-0.013323604129254818,
0.0841403380036354,
-0.13001221418380737,
0.036073360592126846,
0.07543088495731354,
-0.10321200639009476,
0.07379110902547836,
-0.0068534673191607,
0.005426743533462286,
0.006812544539570808,
-0.16181249916553497,
0.05672612413764,
-0.019941296428442,
-0.012209099717438221,
0.017313063144683838,
-0.2010609358549118,
-0.007466768380254507,
-0.046707335859537125,
-0.0589648075401783,
0.009584200568497181,
-0.021017322316765785,
-0.12338412553071976,
0.09723047912120819,
0.001116569503210485,
-0.06670606136322021,
-0.01865183189511299,
0.03669007867574692,
0.09991668909788132,
-0.02648051455616951,
0.1336994618177414,
-0.027254540473222733,
0.07515694200992584,
-0.175013467669487,
-0.00793993379920721,
-0.012013372965157032,
0.03693312406539917,
-0.02674005925655365,
-0.0207707267254591,
0.05938268452882767,
-0.02262757159769535,
0.18207862973213196,
-0.019397424533963203,
0.06667066365480423,
0.05721299722790718,
0.013106883503496647,
0.01558584626764059,
0.08319313079118729,
0.05973837897181511,
0.0017676139250397682,
-0.004349139053374529,
0.031798750162124634,
-0.009058183059096336,
-0.04055638611316681,
-0.16406604647636414,
0.06464677304029465,
0.14954417943954468,
0.051173172891139984,
0.02139807865023613,
0.027088681235909462,
-0.11130707710981369,
-0.08072148263454437,
0.11982464045286179,
-0.01639542542397976,
-0.03426678106188774,
-0.06868444383144379,
0.16933470964431763,
0.14361993968486786,
-0.19647209346294403,
0.0746019259095192,
-0.05006511136889458,
-0.0490703247487545,
-0.1377924680709839,
-0.17217426002025604,
-0.06207772716879845,
-0.048724763095378876,
-0.019698433578014374,
-0.0667402595281601,
0.04839183762669563,
0.05265918746590614,
0.004006898030638695,
-0.012965921312570572,
0.11254379898309708,
0.00356175284832716,
-0.02282772772014141,
0.05048689991235733,
0.06253103911876678,
0.03390306234359741,
-0.09132327884435654,
0.008234100416302681,
-0.000447249214630574,
0.015083888545632362,
0.06469043344259262,
0.018476879224181175,
-0.057288818061351776,
0.018753664568066597,
-0.01892601139843464,
-0.1162450909614563,
0.040920406579971313,
-0.015248272567987442,
-0.0403355173766613,
0.14405398070812225,
0.0321476049721241,
0.009002109058201313,
-0.021963562816381454,
0.23097769916057587,
-0.07918000221252441,
-0.0727902352809906,
-0.15065091848373413,
0.07585040479898453,
-0.06462853401899338,
0.034025877714157104,
0.030585696920752525,
-0.11625860631465912,
0.015242349356412888,
0.16368329524993896,
0.1328435093164444,
-0.009096221067011356,
0.013574071228504181,
0.04220924153923988,
0.005263437982648611,
-0.03252824395895004,
0.019557587802410126,
0.04875345528125763,
0.1523897647857666,
-0.07085705548524857,
0.06430332362651825,
-0.01069768238812685,
-0.08074016869068146,
-0.019284626469016075,
0.10243663191795349,
0.00035568943712860346,
0.00016076915198937058,
-0.07120759785175323,
0.14256687462329865,
-0.08688405156135559,
-0.22196856141090393,
0.06141022965312004,
-0.07179008424282074,
-0.14628954231739044,
-0.046949271112680435,
0.024489209055900574,
-0.01016549114137888,
0.011511257849633694,
0.07719148695468903,
-0.051436688750982285,
0.1672881543636322,
0.04467973858118057,
-0.052464477717876434,
-0.08638637512922287,
0.05645255371928215,
-0.14407914876937866,
0.28069284558296204,
0.020394576713442802,
0.042618103325366974,
0.10270291566848755,
-0.018491527065634727,
-0.1389833390712738,
0.012843544594943523,
0.10804756730794907,
-0.07268691807985306,
0.05566677823662758,
0.1703459918498993,
0.0017549610929563642,
0.13109339773654938,
0.05827563628554344,
-0.05704500898718834,
0.037786830216646194,
-0.09256158769130707,
-0.04868713393807411,
-0.10718972980976105,
0.08463750034570694,
-0.08601583540439606,
0.16170652210712433,
0.12720149755477905,
-0.0676821619272232,
-0.008298310451209545,
-0.019984282553195953,
0.08158870786428452,
0.009103265590965748,
0.10681253671646118,
0.013729413039982319,
-0.18714402616024017,
0.033911801874637604,
0.006192868575453758,
0.10257208347320557,
-0.19699718058109283,
-0.05840226635336876,
0.0409420020878315,
-0.017138797789812088,
-0.07940704375505447,
0.11814602464437485,
0.04777278006076813,
0.03367866948246956,
-0.04106747359037399,
-0.054073747247457504,
0.010431119240820408,
0.14420707523822784,
-0.11052276194095612,
-0.005717637948691845
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# phi-1_5-finetuned-gsm8k
This model is a fine-tuned version of [microsoft/phi-1_5](https://huggingface.co/microsoft/phi-1_5) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 1000
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "mit", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "microsoft/phi-1_5", "model-index": [{"name": "phi-1_5-finetuned-gsm8k", "results": []}]} | null | akashAD/phi-1_5-finetuned-gsm8k | [
"peft",
"tensorboard",
"safetensors",
"phi",
"generated_from_trainer",
"custom_code",
"base_model:microsoft/phi-1_5",
"license:mit",
"region:us"
] | 2024-02-06T07:27:42+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #phi #generated_from_trainer #custom_code #base_model-microsoft/phi-1_5 #license-mit #region-us
|
# phi-1_5-finetuned-gsm8k
This model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 1000
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# phi-1_5-finetuned-gsm8k\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #phi #generated_from_trainer #custom_code #base_model-microsoft/phi-1_5 #license-mit #region-us \n",
"# phi-1_5-finetuned-gsm8k\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
48,
37,
6,
12,
8,
3,
89,
4,
44
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #phi #generated_from_trainer #custom_code #base_model-microsoft/phi-1_5 #license-mit #region-us \n# phi-1_5-finetuned-gsm8k\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.1424432396888733,
0.0838576927781105,
-0.0014683984918519855,
0.08383697271347046,
0.1324382871389389,
0.012549459002912045,
0.11031016707420349,
0.10926586389541626,
-0.10265687108039856,
0.10819031298160553,
0.07279889285564423,
0.04566620662808418,
0.06258460134267807,
0.19331002235412598,
-0.01869063638150692,
-0.21761859953403473,
0.04274963214993477,
-0.014358986169099808,
0.008577966131269932,
0.098378986120224,
0.09014648199081421,
-0.09413911402225494,
0.06309182196855545,
-0.0017950236797332764,
-0.15854115784168243,
-0.030715439468622208,
-0.007648399565368891,
-0.045743778347969055,
0.09567788988351822,
0.005209771450608969,
0.10999830812215805,
-0.0002896619262173772,
0.11998654901981354,
-0.18032881617546082,
0.0013165072305127978,
0.09414782375097275,
0.051819808781147,
0.09837475419044495,
0.07256294786930084,
0.06633251905441284,
0.0646926760673523,
-0.09419376403093338,
0.09156136214733124,
0.04241102561354637,
-0.10789575427770615,
-0.1742057502269745,
-0.11943619698286057,
0.11180554330348969,
0.09464307874441147,
0.12311222404241562,
0.006816181819885969,
0.18692007660865784,
-0.039123814553022385,
0.05862881615757942,
0.21694058179855347,
-0.2725893557071686,
-0.07353249937295914,
0.06132093816995621,
0.0961713120341301,
0.07110760360956192,
-0.10452567040920258,
-0.029958315193653107,
0.05262827128171921,
0.04225026071071625,
0.07401992380619049,
0.008994132280349731,
-0.015394255518913269,
-0.059164226055145264,
-0.1359633207321167,
-0.04124417528510094,
0.13850004971027374,
0.045357946306467056,
-0.06582716107368469,
-0.07999653369188309,
-0.07165341079235077,
-0.17416079342365265,
-0.026863202452659607,
-0.0313502699136734,
0.03169495612382889,
-0.03611092269420624,
-0.0423092395067215,
-0.07477705925703049,
-0.08125030249357224,
-0.09279848635196686,
0.007003446109592915,
0.13499489426612854,
0.052333857864141464,
0.018195295706391335,
-0.01814093627035618,
0.11959482729434967,
-0.024961218237876892,
-0.10712531208992004,
-0.03947441652417183,
-0.006215833127498627,
-0.0276741124689579,
-0.04890032112598419,
-0.033300336450338364,
0.04186642915010452,
0.009743446484208107,
0.15558040142059326,
-0.1544899046421051,
0.05395013466477394,
0.0351061187684536,
0.03717339038848877,
-0.06292091310024261,
0.12189146131277084,
-0.05528358742594719,
0.03159380704164505,
0.027843846008181572,
0.10102835297584534,
0.04919814318418503,
-0.010354852303862572,
-0.09359779208898544,
-0.015002327971160412,
0.12663406133651733,
0.06860102713108063,
-0.013979217037558556,
0.013640285469591618,
-0.05152646824717522,
-0.013302650302648544,
0.09578238427639008,
-0.1025916039943695,
0.05631771683692932,
-0.008954817429184914,
-0.053832340985536575,
-0.032801561057567596,
0.039070457220077515,
-0.0015627631219103932,
-0.05989186465740204,
0.0697157233953476,
-0.09178106486797333,
0.006471419706940651,
-0.09891583770513535,
-0.05996902659535408,
0.03365840017795563,
-0.0735895037651062,
-0.017950408160686493,
-0.11095141619443893,
-0.18177515268325806,
-0.0497233085334301,
-0.004513570107519627,
-0.06713654100894928,
-0.05195300281047821,
0.011079471558332443,
-0.11853552609682083,
0.0016077968757599592,
-0.038558479398489,
0.11503022164106369,
-0.05031760409474373,
0.09868794679641724,
0.02134467475116253,
0.006122797727584839,
-0.03341090306639671,
0.027167025953531265,
-0.061598967760801315,
0.029095778241753578,
-0.15460382401943207,
0.02637113630771637,
-0.09513120353221893,
0.01637106016278267,
-0.11275623738765717,
-0.10043193399906158,
-0.019900353625416756,
-0.025360362604260445,
0.09944945573806763,
0.12511999905109406,
-0.1249527707695961,
-0.013535886071622372,
0.1453157663345337,
-0.10624910145998001,
-0.08532917499542236,
0.06913776695728302,
0.009918972849845886,
0.006362130865454674,
0.03985857218503952,
0.13776427507400513,
0.055127620697021484,
-0.20507846772670746,
-0.004585542716085911,
0.027842532843351364,
0.0691622942686081,
-0.05178147181868553,
0.07180517166852951,
-0.05240415409207344,
0.058251913636922836,
0.03001387044787407,
-0.07604999840259552,
0.0008706315420567989,
-0.09266942739486694,
-0.05719780549407005,
-0.06939785182476044,
-0.08281198889017105,
0.01804242841899395,
0.05706879496574402,
0.02907141111791134,
-0.05216987803578377,
-0.11226148903369904,
0.11974632740020752,
0.14382141828536987,
-0.02769477292895317,
0.027317289263010025,
-0.08375104516744614,
0.11661173403263092,
-0.05600370466709137,
-0.03510730713605881,
-0.2129879742860794,
-0.1253507137298584,
0.034171443432569504,
-0.060282304883003235,
-0.006674350239336491,
-0.015563162043690681,
0.05249738320708275,
0.09479008615016937,
-0.05472433939576149,
-0.03951450064778328,
-0.1334223747253418,
0.008194031193852425,
-0.1399279534816742,
-0.17885121703147888,
-0.074214406311512,
-0.015526960603892803,
0.15972311794757843,
-0.2025233656167984,
0.027999280020594597,
0.009076246060431004,
0.1522817313671112,
0.021143388003110886,
-0.07526854425668716,
-0.009570285677909851,
0.07969944179058075,
-0.0028445885982364416,
-0.10462893545627594,
0.05588749423623085,
0.032575368881225586,
-0.047368232160806656,
-0.07712289690971375,
-0.1668543666601181,
0.10104293376207352,
0.11729353666305542,
0.036901336163282394,
-0.09639877825975418,
0.012295576743781567,
-0.07721798121929169,
-0.03381005674600601,
-0.07688245177268982,
-0.01283374521881342,
0.11602424085140228,
0.029613137245178223,
0.13509507477283478,
-0.07909780740737915,
-0.05475225672125816,
0.030607830733060837,
-0.013123340904712677,
0.00902621727436781,
0.07026781886816025,
0.07086700946092606,
-0.029556892812252045,
0.11433257162570953,
0.08809162676334381,
-0.05392518267035484,
0.08010327816009521,
-0.07606271654367447,
-0.10072554647922516,
-0.016533685848116875,
0.031143542379140854,
0.01948021911084652,
0.16633202135562897,
-0.041841305792331696,
0.021706409752368927,
0.025930440053343773,
0.011444727890193462,
0.04608619213104248,
-0.21135367453098297,
-0.0039657847955822945,
-0.011210652068257332,
-0.036750681698322296,
0.010634359903633595,
-0.026896394789218903,
0.014008719474077225,
0.08446665853261948,
0.03539764881134033,
-0.01579936407506466,
0.01735941506922245,
0.01119672879576683,
-0.09175574779510498,
0.1904529333114624,
-0.09785058349370956,
-0.12326320260763168,
-0.1359362155199051,
0.09688983857631683,
-0.04869198799133301,
-0.020290100947022438,
0.012762323021888733,
-0.05280819535255432,
-0.0435519814491272,
-0.10249630361795425,
-0.016386235132813454,
-0.03541416674852371,
-0.014016684144735336,
0.014883584342896938,
0.025078127160668373,
0.12349973618984222,
-0.11125943064689636,
0.00261088740080595,
-0.0193144753575325,
-0.08785323053598404,
-0.02061428688466549,
0.049819719046354294,
0.10780370235443115,
0.12712174654006958,
0.00015362445265054703,
0.001380203990265727,
-0.03704370930790901,
0.233550563454628,
-0.09862387180328369,
-0.013836166821420193,
0.17638351023197174,
0.024573037400841713,
0.04625020548701286,
0.08443184196949005,
0.034607041627168655,
-0.09202247858047485,
0.0034956582821905613,
0.04175601154565811,
-0.012989235110580921,
-0.2280493527650833,
-0.061287909746170044,
-0.03070315159857273,
-0.032734811305999756,
0.07015925645828247,
0.06806152313947678,
0.061886560171842575,
0.040860503911972046,
-0.03698689118027687,
0.03373357653617859,
-0.01420124527066946,
0.10671305656433105,
0.07519964128732681,
0.032247044146060944,
0.09359011799097061,
-0.03678736463189125,
0.01912551373243332,
0.06740475445985794,
0.014194193296134472,
0.22484590113162994,
-0.006771663203835487,
0.06762414425611496,
0.04648401960730553,
0.1653231829404831,
-0.007900114171206951,
0.02677409164607525,
0.05719996988773346,
0.0027470237109810114,
0.02156653255224228,
-0.07039107382297516,
-0.033160749822854996,
0.013801680877804756,
-0.0644037052989006,
0.05335302650928497,
-0.11241880804300308,
-0.004078080877661705,
0.007034271024167538,
0.26809054613113403,
0.05575239285826683,
-0.32482051849365234,
-0.11378709226846695,
-0.002835772931575775,
-0.01993410289287567,
-0.07462042570114136,
-0.001610980019904673,
0.09748908877372742,
-0.13682733476161957,
0.04258132725954056,
-0.051880016922950745,
0.09370213001966476,
-0.0554274246096611,
0.017181668430566788,
0.0618290901184082,
0.12381858378648758,
0.020525455474853516,
0.0478755347430706,
-0.2119361311197281,
0.2297694981098175,
0.02371714636683464,
0.14189539849758148,
-0.019534198567271233,
0.025447625666856766,
-0.00014947535237297416,
0.10737065970897675,
0.09782364964485168,
-0.01289280317723751,
0.017755214124917984,
-0.21069708466529846,
-0.11360997706651688,
0.008034716360270977,
0.08983559906482697,
-0.01948026567697525,
0.07324542850255966,
-0.044246964156627655,
0.022344378754496574,
0.029218854382634163,
-0.05846625939011574,
-0.18186916410923004,
-0.06548647582530975,
0.03090016171336174,
-0.010136149823665619,
-0.005442182533442974,
-0.10833550989627838,
-0.09965340048074722,
-0.04019661247730255,
0.13859261572360992,
-0.07269393652677536,
-0.06066156178712845,
-0.1585966795682907,
0.09809372574090958,
0.10389606654644012,
-0.053106434643268585,
0.03443416208028793,
-0.0058359126560389996,
0.13290442526340485,
0.025684194639325142,
-0.05283826217055321,
0.06833640486001968,
-0.06562317162752151,
-0.2344721555709839,
-0.04385673254728317,
0.12365970760583878,
0.02203145995736122,
0.04077194258570671,
-0.017931070178747177,
0.028085164725780487,
0.009590750560164452,
-0.10474469512701035,
0.012145580723881721,
0.11044206470251083,
0.0793563574552536,
0.004426788073033094,
-0.054166555404663086,
0.058775924146175385,
-0.019406868144869804,
-0.03401776775717735,
0.05967884138226509,
0.25184252858161926,
-0.08998669683933258,
0.0074303168803453445,
0.07155454158782959,
-0.06990999728441238,
-0.20220418274402618,
0.05017675459384918,
0.09697830677032471,
0.006523967254906893,
-0.004855876788496971,
-0.14469945430755615,
0.07271331548690796,
0.151075541973114,
-0.06181962043046951,
0.07526115328073502,
-0.3007022738456726,
-0.15069317817687988,
0.0714106485247612,
0.13420577347278595,
0.021169055253267288,
-0.17791694402694702,
-0.06169386953115463,
-0.02731735073029995,
-0.12115701287984848,
0.039268817752599716,
-0.15522918105125427,
0.09920886158943176,
-0.004683764651417732,
0.04304852709174156,
0.005574622191488743,
-0.04195231944322586,
0.15982645750045776,
-0.007217908278107643,
0.07973996549844742,
-0.043471939861774445,
0.045999642461538315,
0.082257479429245,
-0.06206245347857475,
0.04895367845892906,
0.04211493209004402,
0.07503030449151993,
-0.10534019023180008,
-0.01171706710010767,
-0.0760820135474205,
0.0636548101902008,
-0.06433825194835663,
-0.04735545814037323,
-0.03842660039663315,
0.06019686162471771,
-0.01788783259689808,
-0.034463070333004,
0.08788397163152695,
0.011912615969777107,
0.14569762349128723,
0.13398544490337372,
0.07502581179141998,
-0.006108473986387253,
-0.1314043253660202,
0.01894509047269821,
-0.05409681797027588,
0.07281853258609772,
-0.11023169010877609,
-0.017378386110067368,
0.10557696968317032,
0.033641643822193146,
0.08651868253946304,
0.045485589653253555,
-0.08389042317867279,
0.009833697229623795,
0.036094196140766144,
-0.11617761105298996,
-0.17970097064971924,
0.004968635737895966,
0.020458310842514038,
-0.10859407484531403,
0.06590739637613297,
0.12489479780197144,
-0.08862049132585526,
-0.014800703153014183,
-0.003311910666525364,
0.01258570235222578,
-0.031330328434705734,
0.18755657970905304,
0.07295694202184677,
0.07022571563720703,
-0.08561529219150543,
0.13256216049194336,
0.07112089544534683,
-0.061593979597091675,
0.04022381454706192,
0.07014662772417068,
-0.11165943741798401,
-0.0287544596940279,
0.07625576853752136,
0.10479254275560379,
-0.03575902804732323,
-0.07951551675796509,
-0.08848448842763901,
-0.09677718579769135,
0.04681235924363136,
0.1092495396733284,
0.04669203609228134,
0.00727305980399251,
0.009341700002551079,
0.043730296194553375,
-0.14416275918483734,
0.06537316739559174,
-0.003120454028248787,
0.08395759761333466,
-0.18060065805912018,
0.12876074016094208,
0.026670241728425026,
0.0282478928565979,
-0.02509334683418274,
0.04181155189871788,
-0.10850849747657776,
-0.013099756091833115,
-0.10192689299583435,
-0.002385158557444811,
-0.02663307636976242,
-0.01548678707331419,
-0.015253352001309395,
-0.06020835041999817,
-0.021278798580169678,
0.07012863457202911,
-0.06071934849023819,
-0.071112260222435,
0.008107118308544159,
0.02875220589339733,
-0.11364510655403137,
0.028363868594169617,
0.0020984397269785404,
-0.09148949384689331,
0.07530727982521057,
0.0291754137724638,
0.056787874549627304,
0.02528381533920765,
-0.0697852373123169,
0.02582360990345478,
0.03542517125606537,
0.009569980204105377,
0.05090307071805,
-0.05087319016456604,
-0.02084122784435749,
-0.02139727771282196,
0.05183222144842148,
0.01927901618182659,
0.08409753441810608,
-0.1364605724811554,
-0.03460657596588135,
-0.030598200857639313,
-0.024866614490747452,
-0.06453864276409149,
0.03420024365186691,
0.12058741599321365,
0.02295711450278759,
0.12388427555561066,
-0.0832090675830841,
0.026478607207536697,
-0.2057638317346573,
-0.010577939450740814,
-0.014360582455992699,
-0.024813905358314514,
-0.07078924030065536,
-0.019399307668209076,
0.10315227508544922,
-0.02783352881669998,
0.08817826956510544,
-0.005755085498094559,
0.15353550016880035,
0.05123801529407501,
-0.03956623002886772,
-0.008678569458425045,
0.03832775354385376,
0.16183257102966309,
0.06659999489784241,
0.007584048435091972,
0.09790511429309845,
0.002973055001348257,
0.045297347009181976,
0.04004504904150963,
0.19336088001728058,
0.16438259184360504,
-0.01909092254936695,
0.09728515893220901,
0.07121063768863678,
-0.09923713654279709,
-0.15426619350910187,
0.028584342449903488,
-0.038555048406124115,
0.10209065675735474,
-0.08613316714763641,
0.11782554537057877,
0.09998346120119095,
-0.1552988737821579,
0.015672700479626656,
-0.05418616905808449,
-0.0865335538983345,
-0.11827187240123749,
0.0014321538619697094,
-0.06749366223812103,
-0.12973268330097198,
0.008788683451712132,
-0.10348277539014816,
-0.010613651014864445,
0.11926499009132385,
-0.0056250980123877525,
-0.003537836018949747,
0.15185628831386566,
0.022139525040984154,
0.00872823130339384,
0.05268701910972595,
0.02219691313803196,
0.018250787630677223,
-0.1056416779756546,
-0.08277220278978348,
0.04948580637574196,
0.001265806844457984,
0.0981825739145279,
-0.018831491470336914,
-0.032680340111255646,
0.053838830441236496,
0.015679731965065002,
-0.0789022445678711,
0.04316197708249092,
0.03092597797513008,
0.015157364308834076,
0.0941469743847847,
0.025347691029310226,
0.004733688198029995,
-0.040713053196668625,
0.2602657675743103,
-0.07871688902378082,
-0.0874863713979721,
-0.11690971255302429,
0.2504158914089203,
0.009109312668442726,
-0.03233812749385834,
0.042613908648490906,
-0.13114362955093384,
-0.01853076182305813,
0.18486745655536652,
0.11822716891765594,
-0.056163981556892395,
-0.022932380437850952,
0.03443508222699165,
-0.026402214542031288,
-0.08429460972547531,
0.13982723653316498,
0.10465177148580551,
0.07590355724096298,
-0.07024034857749939,
0.009576303884387016,
0.010717826895415783,
-0.03577505424618721,
-0.09324167668819427,
0.04289120435714722,
-0.00023835396859794855,
0.011184698902070522,
-0.06677019596099854,
0.06970029324293137,
-0.010397743433713913,
-0.14425049722194672,
0.07762596011161804,
-0.14187464118003845,
-0.18073271214962006,
0.0002834302722476423,
0.0776778906583786,
-0.03299165517091751,
0.0567137748003006,
-0.015952371060848236,
0.02708621881902218,
0.12750166654586792,
-0.03823123872280121,
-0.03678717091679573,
-0.10363809764385223,
0.0616736114025116,
-0.048156432807445526,
0.22665467858314514,
-0.02622438222169876,
0.07589893788099289,
0.11418461799621582,
0.0030821082182228565,
-0.17559652030467987,
0.060599081218242645,
0.0520922876894474,
-0.08605543524026871,
0.02802661433815956,
0.11879880726337433,
-0.022329257801175117,
0.03994784131646156,
0.03025761991739273,
-0.1619497835636139,
-0.03140335902571678,
-0.020583223551511765,
-0.027707943692803383,
-0.06453585624694824,
-0.008739323355257511,
-0.07415542006492615,
0.148212730884552,
0.1545359343290329,
-0.053165145218372345,
-0.012176897376775742,
-0.05987468734383583,
0.04339796304702759,
0.024110883474349976,
0.06353168189525604,
0.007738211657851934,
-0.21638771891593933,
0.03784765675663948,
0.01479805912822485,
0.011663933284580708,
-0.2150687724351883,
-0.06015542522072792,
0.046610359102487564,
-0.05064210295677185,
-0.0917847603559494,
0.0930890142917633,
0.03998000919818878,
0.02520141936838627,
-0.04627702385187149,
-0.12723791599273682,
-0.0442362055182457,
0.1511402130126953,
-0.1345471441745758,
-0.07882246375083923
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | Karajan42/NeuralMiria-Mistral-7B | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T07:28:15+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05921921506524086,
0.15253323316574097,
-0.004925556480884552,
0.01970141939818859,
0.09812989830970764,
0.008722675032913685,
0.07155127823352814,
0.11091651022434235,
-0.02038503810763359,
0.11541511863470078,
0.03161177039146423,
0.09504877775907516,
0.11244720220565796,
0.1593349277973175,
0.0006018498679623008,
-0.22924894094467163,
0.050943523645401,
-0.12565383315086365,
-0.028005311265587807,
0.1202453151345253,
0.14323006570339203,
-0.10873830318450928,
0.07482945919036865,
-0.03924073651432991,
-0.006830108352005482,
-0.03327549248933792,
-0.06254202127456665,
-0.05196645110845566,
0.05287102237343788,
0.06693000346422195,
0.07382122427225113,
0.0121690658852458,
0.09054198116064072,
-0.27071383595466614,
0.02402324043214321,
0.07869837433099747,
-0.00047617589007131755,
0.07642106711864471,
0.049837369471788406,
-0.08698169887065887,
0.07614438980817795,
-0.060363397002220154,
0.14962489902973175,
0.07956483215093613,
-0.09049813449382782,
-0.19196605682373047,
-0.07841940224170685,
0.10002946108579636,
0.18888257443904877,
0.05783533677458763,
-0.02747977338731289,
0.11718999594449997,
-0.08618196099996567,
0.013946855440735817,
0.06651762872934341,
-0.05830651894211769,
-0.055825375020504,
0.07012750208377838,
0.08251979202032089,
0.08537944406270981,
-0.13050076365470886,
-0.011774240992963314,
0.015172234736382961,
0.00940374843776226,
0.0883294939994812,
0.017624128609895706,
0.13745273649692535,
0.04126768559217453,
-0.1351923644542694,
-0.04287068545818329,
0.09870852530002594,
0.035997726023197174,
-0.04835180938243866,
-0.24833782017230988,
-0.023138362914323807,
-0.039952121675014496,
-0.03223174810409546,
-0.0381147637963295,
0.04236193001270294,
-0.01381280180066824,
0.07635250687599182,
-0.0030598659068346024,
-0.08292017132043839,
-0.042900193482637405,
0.07140932232141495,
0.06195797771215439,
0.025352943688631058,
-0.016651969403028488,
0.0064301020465791225,
0.12258180975914001,
0.11147689074277878,
-0.12772345542907715,
-0.053019966930150986,
-0.06414514780044556,
-0.08524893969297409,
-0.04640465974807739,
0.03045455552637577,
0.03743596002459526,
0.047410931438207626,
0.2386423945426941,
0.0032438088674098253,
0.054757438600063324,
0.046099163591861725,
0.014072372578084469,
0.06632840633392334,
0.10764557868242264,
-0.05884917825460434,
-0.09735266119241714,
-0.030795203521847725,
0.10186740756034851,
0.006704956758767366,
-0.041407015174627304,
-0.05594591051340103,
0.06964502483606339,
0.020676078274846077,
0.1224241703748703,
0.07868597656488419,
0.002938423305749893,
-0.07543925195932388,
-0.06281042098999023,
0.18152743577957153,
-0.1571107804775238,
0.0444292388856411,
0.03200872242450714,
-0.03442244604229927,
-0.009351148270070553,
0.00990392453968525,
0.02681080251932144,
-0.02011663094162941,
0.09737543761730194,
-0.05644093081355095,
-0.033681318163871765,
-0.11296935379505157,
-0.0371013842523098,
0.030811145901679993,
0.01213210541754961,
-0.029025491327047348,
-0.0342867337167263,
-0.0882277637720108,
-0.0636090338230133,
0.09107700735330582,
-0.07191670686006546,
-0.04744245857000351,
-0.017612621188163757,
-0.07794062048196793,
0.022423118352890015,
0.017721612006425858,
0.09050743281841278,
-0.021899394690990448,
0.03913994878530502,
-0.056751471012830734,
0.06101011112332344,
0.11571475863456726,
0.028108863160014153,
-0.058606795966625214,
0.06155762821435928,
-0.2421950101852417,
0.10317995399236679,
-0.07758963108062744,
0.051325954496860504,
-0.1530446857213974,
-0.026070065796375275,
0.03956404700875282,
0.012061306275427341,
-0.008345595560967922,
0.1417774260044098,
-0.2185831218957901,
-0.03138069063425064,
0.1676056981086731,
-0.10102425515651703,
-0.07971794903278351,
0.06269615143537521,
-0.05407082289457321,
0.11134804040193558,
0.04596652463078499,
-0.023191405460238457,
0.05842197686433792,
-0.14511504769325256,
-0.00791724119335413,
-0.04188765957951546,
-0.017894908785820007,
0.16635635495185852,
0.07102048397064209,
-0.06073606386780739,
0.07092984020709991,
0.019934939220547676,
-0.016795052215456963,
-0.04869792237877846,
-0.028511613607406616,
-0.10498060286045074,
0.011810078285634518,
-0.059134796261787415,
0.02167343720793724,
-0.021296551451086998,
-0.09382132440805435,
-0.029188871383666992,
-0.17379464209079742,
-0.0012200147612020373,
0.08734307438135147,
-0.010546354576945305,
-0.02201107330620289,
-0.11164727807044983,
0.008580547757446766,
0.03398929536342621,
0.0007392297266051173,
-0.13708379864692688,
-0.059298936277627945,
0.02737307921051979,
-0.16233380138874054,
0.02912268228828907,
-0.05535917729139328,
0.046022266149520874,
0.040077272802591324,
-0.03548351675271988,
-0.0344831608235836,
0.01168955210596323,
0.011000183410942554,
-0.01812567003071308,
-0.25495970249176025,
-0.017501724883913994,
-0.02502158097922802,
0.17353887856006622,
-0.22721131145954132,
0.04271984100341797,
0.07614967226982117,
0.14550280570983887,
0.0073052942752838135,
-0.034482456743717194,
0.014565827324986458,
-0.07198352366685867,
-0.03167816624045372,
-0.06257235258817673,
-0.010083765722811222,
-0.03872835263609886,
-0.06014038994908333,
0.04782424867153168,
-0.16939696669578552,
-0.03236479312181473,
0.10534932464361191,
0.06398996710777283,
-0.14835967123508453,
-0.030286256223917007,
-0.0393594354391098,
-0.047035153955221176,
-0.06618485599756241,
-0.054856978356838226,
0.12015452980995178,
0.05620792135596275,
0.04745647683739662,
-0.07151947915554047,
-0.07490099221467972,
0.007241961546242237,
-0.019977761432528496,
-0.0163256898522377,
0.09354335069656372,
0.06967450678348541,
-0.12794628739356995,
0.09154868870973587,
0.0982460081577301,
0.08392132818698883,
0.10398648679256439,
-0.015390566550195217,
-0.08757331967353821,
-0.041474130004644394,
0.023933125659823418,
0.014664852991700172,
0.1483616679906845,
-0.016296299174427986,
0.054420776665210724,
0.0360836423933506,
-0.013510678894817829,
0.01076538860797882,
-0.09628108888864517,
0.02706051431596279,
0.02971329540014267,
-0.015405743382871151,
0.03466423228383064,
-0.04367179423570633,
0.019455796107649803,
0.09001301974058151,
0.041830018162727356,
0.0396038182079792,
0.010561688803136349,
-0.04398298263549805,
-0.11032342165708542,
0.17876994609832764,
-0.12373854219913483,
-0.2460412234067917,
-0.13813963532447815,
0.010937176644802094,
0.04738753288984299,
-0.011057097464799881,
0.006951550021767616,
-0.06640941649675369,
-0.1170244961977005,
-0.09733203053474426,
0.01991088129580021,
0.04529648274183273,
-0.07728998363018036,
-0.06572148203849792,
0.06318122148513794,
0.037644270807504654,
-0.13899093866348267,
0.023945696651935577,
0.0469096377491951,
-0.0813174769282341,
-0.0011905812425538898,
0.07709334045648575,
0.06798645853996277,
0.17623907327651978,
0.014159789308905602,
-0.023712651804089546,
0.025652561336755753,
0.21002908051013947,
-0.14298869669437408,
0.1094568595290184,
0.1327279806137085,
-0.08898334950208664,
0.08212688565254211,
0.20222385227680206,
0.0385010726749897,
-0.10506977140903473,
0.03657889738678932,
0.027060477063059807,
-0.02792542427778244,
-0.24959829449653625,
-0.06908850371837616,
0.001758498721756041,
-0.053698375821113586,
0.06916391849517822,
0.08716317266225815,
0.09721273928880692,
0.016790922731161118,
-0.10066783428192139,
-0.0790279284119606,
0.05001477152109146,
0.10897587984800339,
-0.001458899350836873,
-0.014394176192581654,
0.09075857698917389,
-0.02953648567199707,
0.01689162664115429,
0.09213569760322571,
0.0019032615236938,
0.1793205291032791,
0.052213337272405624,
0.17340974509716034,
0.07910763472318649,
0.06269825994968414,
0.021207094192504883,
0.006816241890192032,
0.02095629647374153,
0.01695442944765091,
-0.004212336614727974,
-0.0863528773188591,
-0.0027415938675403595,
0.1203664243221283,
0.050876569002866745,
0.03059028834104538,
0.014285655692219734,
-0.03054206818342209,
0.08466528356075287,
0.177787184715271,
0.001063879462890327,
-0.1876421719789505,
-0.07282958924770355,
0.07934894412755966,
-0.08512143790721893,
-0.10675539821386337,
-0.029639042913913727,
0.040873926132917404,
-0.17292065918445587,
0.01861744187772274,
-0.020119842141866684,
0.10806277394294739,
-0.12885749340057373,
-0.017452897503972054,
0.055447377264499664,
0.06997017562389374,
-0.009931124746799469,
0.06633757054805756,
-0.1625119000673294,
0.1177479475736618,
0.01653103344142437,
0.06594116985797882,
-0.09538834542036057,
0.095417320728302,
-0.006962447427213192,
0.007516060955822468,
0.1403670459985733,
0.010755252093076706,
-0.0641925036907196,
-0.0961010679602623,
-0.10299893468618393,
-0.010606445372104645,
0.1309773176908493,
-0.14660196006298065,
0.08697716891765594,
-0.02743646875023842,
-0.0437387153506279,
0.0037594304885715246,
-0.12246467173099518,
-0.13224415481090546,
-0.18235477805137634,
0.05769521743059158,
-0.13171130418777466,
0.040173836052417755,
-0.1089821308851242,
-0.04585907980799675,
-0.021465247496962547,
0.1977471560239792,
-0.23280778527259827,
-0.06815840303897858,
-0.15394872426986694,
-0.08265888690948486,
0.1454220414161682,
-0.04706942290067673,
0.08337214589118958,
0.000301246385788545,
0.19080647826194763,
0.020952312275767326,
-0.017133628949522972,
0.1067209243774414,
-0.09975022822618484,
-0.20161914825439453,
-0.09120959788560867,
0.15868841111660004,
0.13963958621025085,
0.038726504892110825,
-0.004869744647294283,
0.032236017286777496,
-0.021885421127080917,
-0.12115032970905304,
0.02010788396000862,
0.17255425453186035,
0.08749033510684967,
0.026468761265277863,
-0.028463367372751236,
-0.11846643686294556,
-0.07225121557712555,
-0.03745346516370773,
0.02470988966524601,
0.1813775599002838,
-0.07139390707015991,
0.18551595509052277,
0.14274363219738007,
-0.054879751056432724,
-0.19840270280838013,
0.02148755080997944,
0.04472679644823074,
0.0060237692669034,
0.03174281120300293,
-0.20237314701080322,
0.09144619107246399,
0.0006281035020947456,
-0.05034751072525978,
0.13383205235004425,
-0.18327344954013824,
-0.15106844902038574,
0.061150215566158295,
0.04303572699427605,
-0.19199669361114502,
-0.1237611323595047,
-0.08872545510530472,
-0.046805474907159805,
-0.1568751484155655,
0.1029038056731224,
0.0011325168889015913,
0.007591354660689831,
0.03782656043767929,
0.024313677102327347,
0.012553532607853413,
-0.041947584599256516,
0.19289998710155487,
-0.02507353574037552,
0.034427378326654434,
-0.0793621614575386,
-0.06381990760564804,
0.06411149352788925,
-0.057697590440511703,
0.0750909373164177,
-0.025500034913420677,
0.015388053841888905,
-0.10115842521190643,
-0.047956179827451706,
-0.029484452679753304,
0.01986371912062168,
-0.09421123564243317,
-0.09366033226251602,
-0.04838487133383751,
0.0944879949092865,
0.08926530182361603,
-0.037268105894327164,
-0.033034052699804306,
-0.07874293625354767,
0.04173892363905907,
0.17448031902313232,
0.18235735595226288,
0.045147113502025604,
-0.07717937231063843,
-0.0013610349269583821,
-0.014655699953436852,
0.04845907539129257,
-0.22060799598693848,
0.06062275543808937,
0.045259539037942886,
0.01552091259509325,
0.11744016408920288,
-0.020618194714188576,
-0.1619492471218109,
-0.0666290745139122,
0.06087447330355644,
-0.06730270385742188,
-0.1811886727809906,
0.00352504407055676,
0.0753183513879776,
-0.16591353714466095,
-0.03711319714784622,
0.04232833534479141,
-0.011535273864865303,
-0.04050648957490921,
0.013207654468715191,
0.08094717562198639,
0.0073035703971982,
0.07697968184947968,
0.05389590561389923,
0.09186159074306488,
-0.10275198519229889,
0.07336891442537308,
0.08092255145311356,
-0.08580191433429718,
0.029650582000613213,
0.0956844761967659,
-0.0660475566983223,
-0.03553546592593193,
0.039692267775535583,
0.08463539928197861,
0.025261107832193375,
-0.04666709899902344,
0.003693421371281147,
-0.09922701120376587,
0.05857077240943909,
0.11215036362409592,
0.035282451659440994,
0.011146705597639084,
0.03799959644675255,
0.04474346339702606,
-0.07786709815263748,
0.11944296956062317,
0.024733934551477432,
0.020655835047364235,
-0.04009570553898811,
-0.040743377059698105,
0.03469119220972061,
-0.027051862329244614,
-0.011984582990407944,
-0.035381630063056946,
-0.07329677045345306,
-0.014250458218157291,
-0.16089624166488647,
-0.006425157655030489,
-0.039050452411174774,
0.006492188666015863,
0.0227071400731802,
-0.03757927939295769,
0.008156952448189259,
0.012379756197333336,
-0.06891508400440216,
-0.05483170598745346,
-0.0225595161318779,
0.09499263763427734,
-0.16361327469348907,
0.02182857319712639,
0.08322018384933472,
-0.12078364938497543,
0.09284685552120209,
0.016550488770008087,
0.002410374814644456,
0.028476644307374954,
-0.15792103111743927,
0.04754367470741272,
-0.020290223881602287,
0.012727295979857445,
0.04053649678826332,
-0.2180718630552292,
-0.005482743959873915,
-0.04065772518515587,
-0.055209364742040634,
-0.008002875372767448,
-0.03194994851946831,
-0.11256447434425354,
0.09542836248874664,
0.010766619816422462,
-0.0858173593878746,
-0.029525602236390114,
0.032997291535139084,
0.07880192995071411,
-0.02688010409474373,
0.15163032710552216,
-0.004930328112095594,
0.07543973624706268,
-0.17439891397953033,
-0.02280678227543831,
-0.009784235619008541,
0.02145213820040226,
-0.02418927662074566,
-0.016610441729426384,
0.04521343484520912,
-0.027311841025948524,
0.18978725373744965,
-0.02763848751783371,
0.047156915068626404,
0.06419318169355392,
0.01327395811676979,
-0.016141459345817566,
0.11109550297260284,
0.05755641311407089,
0.024413742125034332,
0.02059282548725605,
0.0006552583072334528,
-0.04046328365802765,
-0.012729931622743607,
-0.18779614567756653,
0.06844497472047806,
0.14769941568374634,
0.09005311876535416,
-0.014767808839678764,
0.06981590390205383,
-0.09979446232318878,
-0.11724765598773956,
0.10648569464683533,
-0.06312347948551178,
-0.011802246794104576,
-0.06541955471038818,
0.14070585370063782,
0.1514706313610077,
-0.1892511397600174,
0.06684626638889313,
-0.06704412400722504,
-0.05669668689370155,
-0.11357752978801727,
-0.1923627108335495,
-0.05791294202208519,
-0.05011613294482231,
-0.018368201330304146,
-0.05373769626021385,
0.06899537891149521,
0.057158127427101135,
0.011277895420789719,
0.008883214555680752,
0.0839093029499054,
-0.009658100083470345,
0.001425864058546722,
0.031231271103024483,
0.06669623404741287,
0.016144385561347008,
-0.0304893609136343,
0.01806715875864029,
-0.003015234600752592,
0.033999331295490265,
0.059489116072654724,
0.036065202206373215,
-0.028380198404192924,
0.013694645836949348,
-0.03632815182209015,
-0.11369726806879044,
0.043240632861852646,
-0.028342511504888535,
-0.07773103564977646,
0.13286112248897552,
0.026473212987184525,
0.005609886720776558,
-0.022322779521346092,
0.2495104819536209,
-0.07400858402252197,
-0.09536818414926529,
-0.1448878049850464,
0.11703428626060486,
-0.04134928435087204,
0.06479805707931519,
0.03765689954161644,
-0.10748469084501266,
0.018750222399830818,
0.12525403499603271,
0.1550474315881729,
-0.04537956044077873,
0.019106155261397362,
0.02858782559633255,
0.004584235139191151,
-0.04013598710298538,
0.05142189934849739,
0.06933367252349854,
0.14214643836021423,
-0.05173535272479057,
0.08858583122491837,
0.0017827433766797185,
-0.10212727636098862,
-0.04129546508193016,
0.11294585466384888,
-0.012940747663378716,
0.016553698107600212,
-0.05866444855928421,
0.1253037303686142,
-0.059382375329732895,
-0.23649652302265167,
0.061238259077072144,
-0.07580125331878662,
-0.14206883311271667,
-0.02515989914536476,
0.0734870657324791,
-0.015550101175904274,
0.026368482038378716,
0.07198820263147354,
-0.07507873326539993,
0.18898127973079681,
0.03871531784534454,
-0.05198408663272858,
-0.05836968496441841,
0.07604995369911194,
-0.117560975253582,
0.2752254605293274,
0.01097069587558508,
0.05294901132583618,
0.10413134098052979,
-0.02049596607685089,
-0.13178466260433197,
0.024117950350046158,
0.09550730884075165,
-0.08813395351171494,
0.04131056368350983,
0.21484604477882385,
-0.005940921604633331,
0.1187596246600151,
0.07743308693170547,
-0.07539036870002747,
0.047102998942136765,
-0.1141449362039566,
-0.0771128386259079,
-0.08687382191419601,
0.09549140185117722,
-0.0675748735666275,
0.14216206967830658,
0.12683449685573578,
-0.054658904671669006,
0.010759806260466576,
-0.02898469939827919,
0.045599378645420074,
0.0063186027109622955,
0.10157246887683868,
0.009957551956176758,
-0.18577666580677032,
0.02454824559390545,
0.017152229323983192,
0.10993915796279907,
-0.1806284487247467,
-0.09123970568180084,
0.04470835253596306,
0.0021878182888031006,
-0.06369121372699738,
0.12484876811504364,
0.057084910571575165,
0.04630184918642044,
-0.044473882764577866,
-0.029204387217760086,
-0.0060947248712182045,
0.1420498490333557,
-0.10524781048297882,
-0.003831128589808941
] |
null | null | null |
# Lora of okita_souji_alter/沖田総司〔オルタ〕/冲田总司〔Alter〕 (Fate/Grand Order)
## What Is This?
This is the LoRA model of waifu okita_souji_alter/沖田総司〔オルタ〕/冲田总司〔Alter〕 (Fate/Grand Order).
## How Is It Trained?
* This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion).
* The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs).
* The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest).
* Dataset used for training is the `stage3-p480-800` in [CyberHarem/okita_souji_alter_fgo](https://huggingface.co/datasets/CyberHarem/okita_souji_alter_fgo), which contains 1271 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 1, resolution is 720x720, clustering into 20 buckets.
* Trained for 10000 steps, 40 checkpoints were saved and evaluated.
* **Trigger word is `okita_souji_alter_fgo`.**
* Pruned core tags for this waifu are `dark_skin, dark-skinned_female, white_hair, ahoge, bow, breasts, hair_bow, hair_between_eyes, black_bow, hair_ornament, tassel, large_breasts, long_hair, bangs, very_long_hair, yellow_eyes`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
## How to Use It?
### If You Are Using A1111 WebUI v1.7+
**Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 8250, you need to download [`8250/okita_souji_alter_fgo.pt`](https://huggingface.co/CyberHarem/okita_souji_alter_fgo/resolve/main/8250/okita_souji_alter_fgo.pt) as the embedding and [`8250/okita_souji_alter_fgo.safetensors`](https://huggingface.co/CyberHarem/okita_souji_alter_fgo/resolve/main/8250/okita_souji_alter_fgo.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters.
## Which Step Should I Use?
We selected 5 good steps for you to choose. The best one is step 8250.
1960 images (2.04 GiB) were generated for auto-testing.

The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11).
Here are the preview of the recommended steps:
| Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_1_0 | pattern_1_1 | pattern_2 | pattern_3 | pattern_4_0 | pattern_4_1 | pattern_5 | pattern_6_0 | pattern_6_1 | pattern_7_0 | pattern_7_1 | pattern_8 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 |
|-------:|--------:|:----------|:-------------|:--------------|:----------|:----------------------------------------------------------------------------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:--------------------------------------------|:--------------------------------------------|:--------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:--------------------------------|:------------------------------------|:--------------------------------|:----------------------------------|:----------------------------------------|:----------------------------------------|:----------------------------------------|:------------------------------|:----------------------------------|:----------------------------------|:--------------------------------|:------------------------------------------------|:----------------------------------|:----------------------------------|:------------------------------|:--------------------------------|:--------------------------------------|:--------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------|:--------------------------------------|
| 8250 | 26 | 0.953 | 0.993 | 0.831 | **0.728** | [Download](https://huggingface.co/CyberHarem/okita_souji_alter_fgo/resolve/main/8250/okita_souji_alter_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 7500 | 24 | **0.953** | 0.990 | 0.830 | 0.727 | [Download](https://huggingface.co/CyberHarem/okita_souji_alter_fgo/resolve/main/7500/okita_souji_alter_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 5000 | 16 | 0.951 | **0.994** | 0.828 | 0.720 | [Download](https://huggingface.co/CyberHarem/okita_souji_alter_fgo/resolve/main/5000/okita_souji_alter_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 4000 | 13 | 0.943 | 0.991 | **0.833** | 0.716 | [Download](https://huggingface.co/CyberHarem/okita_souji_alter_fgo/resolve/main/4000/okita_souji_alter_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 9250 | 30 | 0.944 | 0.990 | 0.828 | 0.710 | [Download](https://huggingface.co/CyberHarem/okita_souji_alter_fgo/resolve/main/9250/okita_souji_alter_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
## Anything Else?
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
## All Steps
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* [Steps From 7750 to 10000](all/0.md)
* [Steps From 5250 to 7500](all/1.md)
* [Steps From 2750 to 5000](all/2.md)
* [Steps From 250 to 2500](all/3.md)
| {"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/okita_souji_alter_fgo"], "pipeline_tag": "text-to-image"} | text-to-image | CyberHarem/okita_souji_alter_fgo | [
"art",
"not-for-all-audiences",
"text-to-image",
"dataset:CyberHarem/okita_souji_alter_fgo",
"license:mit",
"region:us"
] | 2024-02-06T07:31:31+00:00 | [] | [] | TAGS
#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/okita_souji_alter_fgo #license-mit #region-us
| Lora of okita\_souji\_alter/沖田総司〔オルタ〕/冲田总司〔Alter〕 (Fate/Grand Order)
====================================================================
What Is This?
-------------
This is the LoRA model of waifu okita\_souji\_alter/沖田総司〔オルタ〕/冲田总司〔Alter〕 (Fate/Grand Order).
How Is It Trained?
------------------
* This model is trained with HCP-Diffusion.
* The auto-training framework is maintained by DeepGHS Team.
* The base model used for training is deepghs/animefull-latest.
* Dataset used for training is the 'stage3-p480-800' in CyberHarem/okita\_souji\_alter\_fgo, which contains 1271 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 1, resolution is 720x720, clustering into 20 buckets.
* Trained for 10000 steps, 40 checkpoints were saved and evaluated.
* Trigger word is 'okita\_souji\_alter\_fgo'.
* Pruned core tags for this waifu are 'dark\_skin, dark-skinned\_female, white\_hair, ahoge, bow, breasts, hair\_bow, hair\_between\_eyes, black\_bow, hair\_ornament, tassel, large\_breasts, long\_hair, bangs, very\_long\_hair, yellow\_eyes'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
How to Use It?
--------------
### If You Are Using A1111 WebUI v1.7+
Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 8250, you need to download '8250/okita\_souji\_alter\_fgo.pt' as the embedding and '8250/okita\_souji\_alter\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.
Which Step Should I Use?
------------------------
We selected 5 good steps for you to choose. The best one is step 8250.
1960 images (2.04 GiB) were generated for auto-testing.
!Metrics Plot
The base model used for generating preview images is Meina/MeinaMix\_V11.
Here are the preview of the recommended steps:
Anything Else?
--------------
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
All Steps
---------
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* Steps From 7750 to 10000
* Steps From 5250 to 7500
* Steps From 2750 to 5000
* Steps From 250 to 2500
| [
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 8250, you need to download '8250/okita\\_souji\\_alter\\_fgo.pt' as the embedding and '8250/okita\\_souji\\_alter\\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 8250.\n\n\n1960 images (2.04 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 7750 to 10000\n* Steps From 5250 to 7500\n* Steps From 2750 to 5000\n* Steps From 250 to 2500"
] | [
"TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/okita_souji_alter_fgo #license-mit #region-us \n",
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 8250, you need to download '8250/okita\\_souji\\_alter\\_fgo.pt' as the embedding and '8250/okita\\_souji\\_alter\\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 8250.\n\n\n1960 images (2.04 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 7750 to 10000\n* Steps From 5250 to 7500\n* Steps From 2750 to 5000\n* Steps From 250 to 2500"
] | [
48,
38,
483
] | [
"passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/okita_souji_alter_fgo #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file."
] | [
0.018004313111305237,
0.005279409233480692,
-0.003870288608595729,
0.08689072728157043,
0.10267359763383865,
0.067413829267025,
0.22239357233047485,
0.08068986237049103,
0.13136602938175201,
-0.057416751980781555,
0.0586223267018795,
0.05877640098333359,
0.0010032362770289183,
0.06141233071684837,
-0.04115862399339676,
-0.16522766649723053,
-0.056865137070417404,
-0.003768856404349208,
0.01446075551211834,
0.012006756849586964,
0.07222204655408859,
0.008455570787191391,
0.09998590499162674,
-0.03898834064602852,
-0.04102565348148346,
0.038139261305332184,
-0.02568133920431137,
-0.05077070742845535,
0.009854903444647789,
0.08378934115171432,
0.09357678145170212,
0.01700710691511631,
0.055910345166921616,
-0.1430647373199463,
0.07068779319524765,
-0.011870957911014557,
-0.11864294111728668,
-0.0022235086653381586,
0.018923059105873108,
-0.006918507162481546,
0.13285018503665924,
0.03908782824873924,
-0.12171832472085953,
0.04349420964717865,
-0.14089687168598175,
-0.021841665729880333,
-0.037726446986198425,
0.04124177247285843,
0.15147486329078674,
0.08310973644256592,
0.00969090685248375,
0.03478749468922615,
-0.06689942628145218,
0.07495424896478653,
0.10461325943470001,
-0.09416712820529938,
-0.07397641241550446,
0.08624247461557388,
0.025661084800958633,
0.14567671716213226,
-0.1123538687825203,
0.07275687158107758,
0.07752847671508789,
-0.0441158302128315,
-0.1622324138879776,
-0.10655280202627182,
-0.2008945196866989,
-0.0095868781208992,
0.0052193584851920605,
0.03341013565659523,
0.4343998432159424,
0.08082453161478043,
0.019942020997405052,
0.07871661335229874,
-0.07603304833173752,
0.012883811257779598,
-0.09104607999324799,
0.13357211649417877,
0.0430692695081234,
0.09879735112190247,
-0.010739143937826157,
-0.10555247962474823,
-0.12687548995018005,
-0.06732011586427689,
-0.08366110920906067,
-0.009025722742080688,
0.008097406476736069,
0.11054704338312149,
-0.19538874924182892,
0.002268069889396429,
-0.10308320075273514,
-0.12482370436191559,
0.008766870014369488,
-0.08942273259162903,
0.1662617176771164,
0.0706154853105545,
-0.0026271201204508543,
0.023757098242640495,
0.2541142702102661,
0.10286720842123032,
0.16795942187309265,
0.04629978537559509,
-0.11403738707304001,
0.1269763857126236,
0.0600099042057991,
-0.07952289283275604,
-0.013324888423085213,
-0.0504189170897007,
0.1450349986553192,
-0.043407127261161804,
0.09868987649679184,
-0.06311258673667908,
-0.10893670469522476,
0.013137614354491234,
-0.11964468657970428,
0.05649259313941002,
0.04642718657851219,
0.009712032042443752,
-0.052421342581510544,
0.04739740863442421,
0.05012999475002289,
-0.03587246313691139,
0.002406142419204116,
-0.010425685904920101,
-0.057150136679410934,
0.016660558059811592,
0.10026618838310242,
0.04326733946800232,
0.06007993593811989,
0.014605944976210594,
-0.027233511209487915,
0.014717569574713707,
-0.0322566032409668,
0.030022546648979187,
0.05231904983520508,
0.026751305907964706,
0.0956505537033081,
-0.16358962655067444,
-0.06584909558296204,
-0.02583387866616249,
0.025601062923669815,
0.01086888276040554,
0.06419738382101059,
-0.00008428659930359572,
0.04472500830888748,
-0.0005273103597573936,
-0.03398794308304787,
0.04197883605957031,
-0.11718638986349106,
0.08858828246593475,
-0.015566031448543072,
0.07236896455287933,
-0.20927299559116364,
0.000006620420208491851,
-0.06596530973911285,
0.012686437927186489,
0.05782926455140114,
-0.023269517347216606,
-0.09586681425571442,
0.09910393506288528,
0.006917658261954784,
0.0759684219956398,
-0.11640870571136475,
0.05789996311068535,
0.009612156078219414,
0.10769382119178772,
-0.07937392592430115,
0.006877587176859379,
0.09195594489574432,
-0.1503421664237976,
-0.15201012790203094,
0.09230823814868927,
-0.013964729383587837,
0.040228601545095444,
0.029359150677919388,
0.16269466280937195,
0.156608447432518,
-0.18001480400562286,
-0.025259431451559067,
0.04759915545582771,
-0.031124653294682503,
-0.08039643615484238,
0.0008394009782932699,
0.09119370579719543,
0.024814894422888756,
0.03967154398560524,
-0.048239633440971375,
0.11883112788200378,
-0.029459068551659584,
-0.0757276713848114,
-0.034584902226924896,
-0.06775323301553726,
-0.0868164673447609,
0.05701629817485809,
-0.017635172232985497,
-0.04523307830095291,
0.017111312597990036,
-0.16295695304870605,
0.1525372713804245,
0.014720721170306206,
0.02110442891716957,
-0.07657250761985779,
0.13788163661956787,
-0.012599700130522251,
0.012234210968017578,
0.008241657167673111,
-0.030657950788736343,
-0.09987516701221466,
0.2237979918718338,
0.09022863209247589,
0.1263902187347412,
0.06200823187828064,
-0.03782343491911888,
-0.06419990956783295,
-0.0001019401679513976,
0.007795699872076511,
-0.03345494717359543,
0.02809380181133747,
-0.08618368953466415,
0.06874196976423264,
-0.01984528824687004,
0.04454059526324272,
-0.00741590978577733,
-0.01745240017771721,
0.086218923330307,
0.022745046764612198,
-0.030320852994918823,
0.09370710700750351,
0.04927536100149155,
-0.017536871135234833,
-0.057300910353660583,
0.01286886166781187,
0.0725032389163971,
-0.0019516441971063614,
-0.08531339466571808,
0.03559568151831627,
-0.013555269688367844,
0.07134474813938141,
0.20576994121074677,
-0.186514213681221,
0.05713041499257088,
0.017016559839248657,
0.03450574725866318,
0.05272049829363823,
-0.0029955932404845953,
-0.010081112384796143,
0.03744341433048248,
-0.03889821469783783,
0.06597084552049637,
-0.018648218363523483,
0.08946961164474487,
-0.01197980996221304,
-0.15393751859664917,
-0.0217975452542305,
-0.0034642752725631,
0.18713197112083435,
-0.16593454778194427,
0.08888313919305801,
0.18254360556602478,
-0.10461027175188065,
0.16136838495731354,
-0.020019594579935074,
-0.00793539546430111,
0.0024432186037302017,
0.038064129650592804,
-0.006067194975912571,
0.09329017996788025,
-0.10784327983856201,
-0.00029328398522920907,
0.02338687889277935,
-0.0803174152970314,
0.03796910122036934,
-0.12741780281066895,
-0.1258125752210617,
-0.07447007298469543,
-0.02848495915532112,
-0.0139699662104249,
0.038080036640167236,
-0.04032735154032707,
0.07597264647483826,
-0.10740358382463455,
-0.06346004456281662,
0.006863175425678492,
-0.06890005618333817,
0.024225512519478798,
0.010586695745587349,
-0.06065406650304794,
-0.14441828429698944,
-0.13821327686309814,
-0.08660480380058289,
-0.16080641746520996,
0.007151975762099028,
0.08933915942907333,
-0.12196493148803711,
-0.0454227551817894,
0.004061737097799778,
-0.04045448452234268,
0.10234321653842926,
-0.07216586917638779,
0.007272614166140556,
0.053604718297719955,
-0.029803622514009476,
-0.14723287522792816,
-0.015548961237072945,
-0.061355557292699814,
-0.08218064904212952,
0.16636618971824646,
-0.13562020659446716,
0.19241614639759064,
-0.05019081383943558,
0.0625682920217514,
0.050626128911972046,
0.05967124179005623,
0.12036032229661942,
-0.11546584218740463,
0.06169433519244194,
0.2158021777868271,
0.04529132321476936,
0.07121805101633072,
0.1263630986213684,
0.0707736685872078,
-0.09951642155647278,
0.03604402393102646,
0.078288733959198,
-0.09840096533298492,
-0.08293944597244263,
-0.048127952963113785,
-0.11576645076274872,
-0.03075915202498436,
0.03800472244620323,
0.08209160715341568,
0.08023890107870102,
0.12008453905582428,
-0.05187280476093292,
-0.01731407456099987,
0.10915163904428482,
0.03873920068144798,
0.06723608821630478,
0.008636199869215488,
0.04622752219438553,
-0.12728212773799896,
-0.047973018139600754,
0.1613427996635437,
0.175424724817276,
0.19383947551250458,
0.023590702563524246,
0.051555290818214417,
0.11345598101615906,
0.08152687549591064,
0.10805690288543701,
0.041064128279685974,
0.04676949232816696,
0.011918236501514912,
-0.07641598582267761,
-0.046564981341362,
0.025671498849987984,
0.017140526324510574,
-0.012075834907591343,
-0.1529841274023056,
0.0797124058008194,
-0.0015484371688216925,
0.07540738582611084,
0.1044805496931076,
0.009546296671032906,
-0.1003250852227211,
0.16658997535705566,
0.09868799895048141,
0.08202031254768372,
-0.07006444782018661,
0.13034367561340332,
0.08182990550994873,
-0.01662665791809559,
0.14448870718479156,
0.04134425148367882,
0.1539786159992218,
-0.0267501138150692,
-0.0642232596874237,
-0.0658060610294342,
-0.06093696504831314,
0.004499631933867931,
0.03578702732920647,
-0.2263437956571579,
0.074039988219738,
0.04915957897901535,
0.01960246078670025,
-0.010602437891066074,
-0.06023623049259186,
0.16742540895938873,
0.16742174327373505,
0.09184038639068604,
0.014545186422765255,
-0.04443568363785744,
-0.027289044111967087,
-0.07250235229730606,
0.04552633315324783,
0.0028435303829610348,
0.03662767633795738,
-0.051497820764780045,
-0.08109772205352783,
-0.015704717487096786,
-0.015497179701924324,
0.0550173819065094,
-0.06367730349302292,
-0.0931914672255516,
-0.04183477535843849,
0.24738965928554535,
-0.009199017658829689,
0.03516454994678497,
0.0630287453532219,
0.04117124155163765,
-0.0373271219432354,
0.05922640860080719,
-0.009050443768501282,
-0.03162610903382301,
-0.029191922396421432,
0.017639897763729095,
-0.0034983642399311066,
-0.06136368587613106,
-0.061948902904987335,
-0.02945270761847496,
-0.10087602585554123,
-0.09265527874231339,
0.007222190033644438,
-0.0302557572722435,
-0.015354445204138756,
-0.01826363615691662,
0.024281807243824005,
-0.07610498368740082,
-0.03597504645586014,
0.0044507491402328014,
0.02389426715672016,
-0.04849950969219208,
-0.13145911693572998,
0.007559818681329489,
0.0063176280818879604,
-0.02707507647573948,
0.03055698424577713,
-0.10478174686431885,
-0.0906139612197876,
-0.06562596559524536,
-0.03970169648528099,
0.07861199975013733,
0.235984668135643,
0.006226572208106518,
-0.008576059713959694,
0.1481829136610031,
-0.0928981602191925,
-0.29455313086509705,
-0.17404553294181824,
-0.1664372831583023,
-0.08906181156635284,
0.021602371707558632,
-0.044297169893980026,
0.036211661994457245,
0.10171344876289368,
-0.04042493924498558,
0.21541962027549744,
-0.14732985198497772,
-0.1024184301495552,
0.0820145383477211,
0.08721398562192917,
0.3140697777271271,
-0.2608256936073303,
0.017820408567786217,
-0.09637767821550369,
-0.04999391734600067,
0.03755008056759834,
-0.09426204115152359,
0.11984794586896896,
0.029445819556713104,
0.03712725266814232,
-0.0001833746791817248,
-0.008963445201516151,
0.1649036407470703,
-0.0852925106883049,
0.1250980794429779,
-0.13250289857387543,
-0.09306475520133972,
0.19650016725063324,
-0.018549615517258644,
-0.0008316179155372083,
-0.19487088918685913,
-0.05210651457309723,
-0.032575685530900955,
0.030852092429995537,
-0.025519808754324913,
0.07582660019397736,
-0.0023626787588000298,
-0.016428600996732712,
-0.1325691044330597,
-0.007605682592839003,
-0.04877253621816635,
0.07324163615703583,
0.2299787998199463,
-0.08356215804815292,
-0.04673490673303604,
0.048175081610679626,
-0.007466231472790241,
0.10564754158258438,
-0.0011251777177676558,
-0.0711270198225975,
-0.03805333375930786,
0.08795233815908432,
-0.19107794761657715,
0.05508917570114136,
0.010100395418703556,
-0.019184459000825882,
0.04828394949436188,
0.00879903044551611,
0.010699015110731125,
0.12126905471086502,
0.1779613494873047,
-0.013580281287431717,
-0.04766722396016121,
-0.03427252173423767,
-0.005685602314770222,
0.1296255737543106,
-0.039985038340091705,
0.10262341052293777,
0.01806926354765892,
0.03588292747735977,
0.0008506966987624764,
0.08396273851394653,
-0.1103527843952179,
-0.09943437576293945,
0.08611385524272919,
-0.043997351080179214,
-0.08697459846735,
0.08381827175617218,
0.05453719198703766,
0.05317367985844612,
-0.007982607930898666,
0.04445471614599228,
0.021485215052962303,
-0.12664099037647247,
0.025010021403431892,
0.21418339014053345,
-0.04952111840248108,
-0.08558912575244904,
-0.07393472641706467,
-0.0020150316413491964,
-0.12236712872982025,
0.046917326748371124,
0.013545331545174122,
-0.022056110203266144,
0.12404441833496094,
-0.055688101798295975,
-0.04637289419770241,
-0.011917537078261375,
-0.04838922247290611,
0.020942244678735733,
-0.14712806046009064,
-0.20257669687271118,
0.04053715988993645,
0.01603989116847515,
-0.06154671311378479,
-0.08657411485910416,
-0.09117992967367172,
0.05730448290705681,
-0.13644516468048096,
0.13454526662826538,
-0.07513263821601868,
0.050517793744802475,
-0.032711245119571686,
-0.0670284554362297,
-0.1105230301618576,
-0.03448401391506195,
-0.0571703277528286,
-0.02531195431947708,
0.06384848058223724,
0.007886257022619247,
-0.10705217719078064,
-0.11811966449022293,
0.07200603932142258,
-0.013402299024164677,
-0.01167761255055666,
0.041454073041677475,
-0.07816992700099945,
0.023623088374733925,
-0.22674228250980377,
-0.04986415058374405,
0.10202659666538239,
0.02719096839427948,
-0.09084676206111908,
0.13244928419589996,
0.04492061957716942,
-0.006497897207736969,
0.02852233499288559,
-0.002513327868655324,
0.1390613466501236,
-0.09417004138231277,
0.00945235788822174,
-0.120817631483078,
-0.1770409643650055,
-0.041549600660800934,
0.027102507650852203,
0.2493157684803009,
0.09218906611204147,
0.12492690980434418,
-0.03982825577259064,
0.04151028022170067,
-0.01669730804860592,
0.06490851938724518,
0.01765197142958641,
-0.10845188796520233,
-0.009767184965312481,
-0.17947918176651,
-0.07020381093025208,
-0.060531437397003174,
0.13452817499637604,
0.050706274807453156,
-0.13860248029232025,
-0.007818679325282574,
0.09120428562164307,
-0.1386479288339615,
-0.003410906996577978,
0.17270900309085846,
-0.048638634383678436,
0.024802183732390404,
-0.14065560698509216,
0.03712792322039604,
0.07507609575986862,
-0.028394587337970734,
-0.014836032874882221,
0.10389664769172668,
0.0021638728212565184,
0.016851412132382393,
0.028666038066148758,
-0.04163854569196701,
0.09140182286500931,
-0.03791545331478119,
0.03787286952137947,
-0.00912556890398264,
-0.048072364181280136,
-0.14302849769592285,
0.18454210460186005,
-0.01331277284771204,
0.004885847680270672,
-0.053664304316043854,
-0.009526096284389496,
-0.10435216873884201,
-0.11254928261041641,
-0.08292610198259354,
-0.1373451054096222,
0.09110214561223984,
-0.05707018822431564,
0.015232672914862633,
0.0023041260428726673,
-0.0002424439589958638,
-0.09772542119026184,
0.01497634220868349,
-0.14791615307331085,
-0.05149274319410324,
0.014437385834753513,
-0.02347622625529766,
-0.01986534520983696,
-0.05950287729501724,
-0.038082804530858994,
0.023908521980047226,
-0.07576607912778854,
-0.06383704394102097,
0.062443215399980545,
0.08443363755941391,
0.08058847486972809,
-0.17429734766483307,
-0.10903822630643845,
-0.06507230550050735,
0.02708740159869194,
0.0741405263543129,
0.1996632218360901,
0.03495791554450989,
-0.01423814706504345,
0.0532553605735302,
0.1335708498954773,
0.032961200922727585,
-0.06000073254108429,
-0.07307877391576767,
-0.12993323802947998,
-0.12634243071079254,
-0.02844991907477379,
-0.051524724811315536,
-0.03858793526887894,
0.015512398444116116,
0.2149534672498703,
0.15246643126010895,
-0.13598909974098206,
0.03676024079322815,
-0.06618587672710419,
0.03120616264641285,
-0.04044143855571747,
0.1328694224357605,
0.04528146982192993,
0.15529212355613708,
-0.03736112639307976,
-0.003804129781201482,
-0.05835461616516113,
0.023530101403594017,
-0.0816308781504631,
0.023762771859765053,
-0.014548109844326973,
-0.06162723898887634,
-0.06066225469112396,
0.08519385010004044,
-0.13372623920440674,
0.041561491787433624,
0.20496608316898346,
-0.12464139610528946,
-0.006628313101828098,
-0.043952781707048416,
0.026174062862992287,
0.11907965689897537,
0.025357084348797798,
-0.08878488838672638,
-0.023479551076889038,
-0.008647874929010868,
0.02528553456068039,
-0.16546274721622467,
-0.09052560478448868,
-0.01810574159026146,
-0.1332714855670929,
0.13311463594436646,
-0.006970756687223911,
0.004847371485084295,
0.02241518720984459,
-0.06821309030056,
-0.0016174546908587217,
0.18586601316928864,
0.008507916703820229,
-0.041341427713632584,
-0.03183380886912346,
-0.05414695665240288,
-0.08879021555185318,
0.043851129710674286,
0.08342686295509338,
0.04203200712800026,
0.0014684846391901374,
0.15053515136241913,
-0.017293037846684456,
-0.041030462831258774,
0.14704012870788574,
-0.2071586400270462,
0.09308270364999771,
-0.01357573177665472,
-0.01264213863760233,
-0.0782143622636795,
-0.053405389189720154,
0.06979024410247803,
0.06619912385940552,
-0.17800799012184143,
-0.02376565895974636,
0.06137804314494133,
-0.09663014858961105,
0.08755842596292496,
0.04018404707312584,
-0.08655590564012527,
0.015718309208750725,
-0.1221441701054573,
-0.015599558129906654,
-0.1046939566731453,
0.028104545548558235,
0.18711307644844055,
-0.039579715579748154,
0.009196686558425426,
-0.12504301965236664,
0.05785565823316574,
-0.031652454286813736,
-0.03493712097406387,
-0.0705406442284584
] |
null | null | stable-baselines3 |
# **ppo** Agent playing **LunarLander-v2**
This is a trained model of a **ppo** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "ppo", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "230.97 +/- 32.20", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Helaly6484/ppo-LunarLander-v2 | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-06T07:33:18+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# ppo Agent playing LunarLander-v2
This is a trained model of a ppo agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# ppo Agent playing LunarLander-v2\nThis is a trained model of a ppo agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# ppo Agent playing LunarLander-v2\nThis is a trained model of a ppo agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# ppo Agent playing LunarLander-v2\nThis is a trained model of a ppo agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.046836115419864655,
0.029148947447538376,
-0.0046101463958621025,
0.0333307646214962,
0.07860215753316879,
-0.03898412734270096,
0.10734876990318298,
0.06503576785326004,
0.09938327223062515,
0.026071077212691307,
0.12705640494823456,
0.06314948946237564,
0.022638585418462753,
0.26038658618927,
0.04753579944372177,
-0.20013634860515594,
0.05630762502551079,
-0.06510073691606522,
0.11960826069116592,
0.07872214913368225,
-0.003378120716661215,
-0.06956975162029266,
0.022916419431567192,
-0.04493887722492218,
0.06570369005203247,
0.04432268068194389,
-0.005315820220857859,
-0.07050011307001114,
0.09956791251897812,
-0.06057116389274597,
0.09341789782047272,
0.04487678036093712,
0.13019762933254242,
-0.18420742452144623,
0.03289811313152313,
0.021116411313414574,
-0.06533198803663254,
0.01074197143316269,
-0.011246061883866787,
0.030265890061855316,
0.14826522767543793,
0.01700606197118759,
0.09245615452528,
0.004316551610827446,
-0.15238945186138153,
-0.0429522879421711,
0.015288358554244041,
0.14146001636981964,
0.08307067304849625,
0.02736084535717964,
0.014232230372726917,
0.17466716468334198,
-0.0598083958029747,
-0.0007140404195524752,
0.19214463233947754,
-0.29843229055404663,
-0.014825547114014626,
0.3281037509441376,
0.04430875554680824,
0.07152628898620605,
-0.06727654486894608,
-0.02267330326139927,
0.017609406262636185,
-0.008723982609808445,
0.006690485868602991,
-0.046117816120386124,
0.15623193979263306,
0.05648185685276985,
-0.05835650488734245,
-0.126860111951828,
0.1257469654083252,
-0.025604767724871635,
0.004147470463067293,
-0.03489384427666664,
-0.005394438747316599,
0.040811534970998764,
-0.024257337674498558,
-0.0450916588306427,
0.03914261981844902,
0.061587441712617874,
0.01780693233013153,
-0.15356993675231934,
-0.10287962108850479,
-0.05435193330049515,
0.010658602230250835,
0.14688551425933838,
-0.002752156462520361,
0.06265412271022797,
-0.002705513732507825,
0.1165628656744957,
-0.08186501264572144,
-0.031962186098098755,
-0.0337538868188858,
-0.08483044058084488,
-0.048444680869579315,
-0.030201584100723267,
-0.03581640124320984,
0.09807024151086807,
0.09538924694061279,
0.06523062288761139,
0.07620259374380112,
0.05250915512442589,
-0.03316764906048775,
0.08620085567235947,
0.027623726055026054,
0.09951308369636536,
0.06621243059635162,
0.007713210768997669,
0.12279512733221054,
0.08414377272129059,
0.057919979095458984,
-0.07475992292165756,
-0.18042846024036407,
0.040299464017152786,
-0.03430815413594246,
0.024281073361635208,
0.09808416664600372,
-0.0516606830060482,
-0.0982866883277893,
-0.03901293873786926,
-0.013795137405395508,
-0.053415097296237946,
-0.026966353878378868,
0.05364292114973068,
-0.007499905303120613,
-0.017810121178627014,
0.022455766797065735,
0.07808789610862732,
0.05275462567806244,
-0.18647731840610504,
0.04211502522230148,
-0.0006324854912236333,
-0.13136817514896393,
0.0023621090222150087,
0.0054101417772471905,
0.048687130212783813,
0.00778453191742301,
-0.11066577583551407,
-0.35198354721069336,
-0.06647075712680817,
0.05136953666806221,
-0.0659802258014679,
-0.16531556844711304,
-0.12089970707893372,
-0.000005948631041974295,
-0.06583793461322784,
-0.054848603904247284,
-0.1268351823091507,
-0.044348638504743576,
0.032961465418338776,
-0.007107375655323267,
0.13101333379745483,
-0.016662850975990295,
-0.0055732931941747665,
-0.1321118175983429,
0.028904471546411514,
-0.23634015023708344,
0.1056433916091919,
-0.05123000964522362,
0.05661416053771973,
-0.0291205495595932,
-0.07231917232275009,
0.06389991194009781,
0.033755429089069366,
-0.0011897480580955744,
0.14667931199073792,
-0.2253977656364441,
-0.013240323401987553,
0.014090153388679028,
-0.16628728806972504,
-0.05289255827665329,
-0.07273607701063156,
-0.06319448351860046,
0.10834063589572906,
0.01703127846121788,
0.14526726305484772,
-0.0654037594795227,
-0.06764110177755356,
0.201731339097023,
-0.005081115756183863,
-0.13724522292613983,
-0.07632984220981598,
0.09570551663637161,
0.0270902868360281,
-0.09137041121721268,
-0.029118968173861504,
-0.2006516009569168,
0.007131924852728844,
-0.037081144750118256,
-0.0626172199845314,
0.013336562551558018,
-0.03789834678173065,
0.18330508470535278,
0.060056064277887344,
0.06028566509485245,
-0.05725764110684395,
0.020750336349010468,
0.13394932448863983,
0.10046449303627014,
0.011613275855779648,
0.04091762378811836,
-0.039767075330019,
0.12293390929698944,
-0.011679338291287422,
-0.02466016821563244,
-0.08827918022871017,
-0.0424991138279438,
-0.03174952045083046,
0.17803902924060822,
0.07580552250146866,
0.20964370667934418,
0.10187480598688126,
0.06859833002090454,
0.02467496693134308,
-0.09892892837524414,
-0.1513243317604065,
-0.017033124342560768,
0.05978599190711975,
-0.1504514217376709,
-0.08873400837182999,
-0.08720777183771133,
-0.12902545928955078,
-0.16854237020015717,
-0.009030645713210106,
-0.20513217151165009,
0.06397808343172073,
0.04221808537840843,
-0.024473631754517555,
0.040763527154922485,
0.09776193648576736,
-0.006766085512936115,
-0.03350764140486717,
0.09113463014364243,
0.0339411236345768,
-0.05894125998020172,
0.002159892814233899,
0.10654745250940323,
0.167886883020401,
0.11892400681972504,
-0.03950030729174614,
-0.00797593966126442,
0.037586651742458344,
-0.00787398125976324,
0.045234404504299164,
0.03564005345106125,
0.174978569149971,
0.19163475930690765,
0.054425641894340515,
0.045060668140649796,
-0.06357864290475845,
0.06746527552604675,
-0.10453670471906662,
-0.13315267860889435,
-0.060281265527009964,
0.07237110286951065,
0.002528173616155982,
0.01673571579158306,
0.0813232809305191,
0.036005157977342606,
0.05432288721203804,
0.17630670964717865,
0.023368367925286293,
-0.09020144492387772,
-0.02028929628431797,
0.01469753310084343,
0.03567739948630333,
0.08054397255182266,
-0.018525345250964165,
-0.034617383033037186,
0.030597951263189316,
-0.10628323256969452,
0.015886828303337097,
-0.13278692960739136,
-0.13511739671230316,
0.04097125679254532,
-0.03287828341126442,
0.003636467969045043,
0.04588380455970764,
0.003995963837951422,
0.03342653438448906,
0.05110451206564903,
-0.07962232828140259,
0.016776930540800095,
-0.03690068796277046,
-0.02884543687105179,
0.025741247460246086,
-0.07139301300048828,
-0.18299360573291779,
-0.19624201953411102,
-0.009616806171834469,
-0.08242443203926086,
0.10203871876001358,
0.0029358756728470325,
-0.11338692158460617,
0.04346148669719696,
-0.08161032199859619,
0.011710702441632748,
-0.029931258410215378,
-0.07436663657426834,
0.08867894858121872,
0.1338450163602829,
-0.023765869438648224,
-0.001976399915292859,
-0.03301612660288811,
-0.11618808656930923,
-0.1620049625635147,
0.07671657204627991,
0.04169827327132225,
-0.005005182232707739,
0.12431041151285172,
0.004772900138050318,
0.0337822288274765,
-0.030233319848775864,
0.020963504910469055,
-0.0011773804435506463,
-0.08059336245059967,
0.29808545112609863,
0.030424362048506737,
-0.002129836706444621,
0.019213907420635223,
0.0015294472686946392,
-0.06707710027694702,
0.05007791519165039,
-0.09769324958324432,
-0.09721370786428452,
-0.0554092638194561,
-0.05823989585042,
-0.09423423558473587,
0.06323466449975967,
0.08048315346240997,
-0.014393172226846218,
-0.08070304989814758,
0.040851231664419174,
0.1025143712759018,
-0.0735236182808876,
-0.05948677286505699,
0.08029131591320038,
0.1200590506196022,
-0.07887734472751617,
0.042977310717105865,
-0.0129091152921319,
0.015458969399333,
0.03176102787256241,
0.06321718543767929,
-0.014324281364679337,
0.06158633902668953,
-0.19901496171951294,
0.05949528515338898,
0.12855088710784912,
0.05994880944490433,
0.07665392756462097,
0.05055674538016319,
-0.08722293376922607,
-0.006576757412403822,
-0.0466182641685009,
-0.2378644049167633,
0.15081806480884552,
0.1389361023902893,
0.075883649289608,
-0.002698163967579603,
-0.004065473098307848,
-0.06226743012666702,
0.09483174979686737,
0.054507236927747726,
-0.15578708052635193,
-0.01970132626593113,
0.08463291823863983,
0.01937125250697136,
0.025867357850074768,
0.08189862221479416,
0.02765529789030552,
-0.192025825381279,
-0.02062750607728958,
0.08543536812067032,
0.01735302060842514,
-0.011789120733737946,
-0.022581497207283974,
-0.18128694593906403,
0.06491454690694809,
0.007404951844364405,
0.0625278577208519,
-0.1622362583875656,
0.1434682458639145,
-0.0969342514872551,
0.03366238251328468,
0.001902480609714985,
-0.019431501626968384,
0.06878815591335297,
-0.07028315961360931,
0.11821826547384262,
0.027626467868685722,
-0.011997345834970474,
-0.1380508989095688,
-0.17975109815597534,
-0.0335431769490242,
-0.13332068920135498,
-0.09446258842945099,
0.05975805222988129,
0.0442298986017704,
0.04146953672170639,
-0.06777696311473846,
0.1665887087583542,
-0.04885237291455269,
0.04810330271720886,
-0.10070829838514328,
-0.05236319079995155,
0.024577336385846138,
-0.12383751571178436,
-0.10444215685129166,
0.04250117763876915,
0.14871153235435486,
0.14602680504322052,
-0.039817605167627335,
-0.026508336886763573,
-0.06403326243162155,
0.012399613857269287,
0.007399255875498056,
0.04305039346218109,
0.024576807394623756,
0.07409482449293137,
-0.0939774140715599,
-0.12641772627830505,
0.007888003252446651,
-0.07329332083463669,
-0.06160348653793335,
-0.019737552851438522,
0.15691520273685455,
0.05116850882768631,
0.09725116938352585,
-0.007945125922560692,
0.055832162499427795,
-0.014031196013092995,
-0.07419904321432114,
0.12231061607599258,
0.0641794428229332,
-0.06671403348445892,
0.08550131320953369,
0.056189171969890594,
0.051594819873571396,
0.03600400686264038,
-0.004640121012926102,
0.2579076290130615,
0.21565678715705872,
-0.02375211752951145,
0.24839943647384644,
0.02127057872712612,
-0.016432587057352066,
-0.17445705831050873,
-0.06797587871551514,
-0.01932806894183159,
0.04260750859975815,
0.059323329478502274,
-0.17399176955223083,
-0.014728029258549213,
-0.053108539432287216,
-0.05106041580438614,
-0.11349649727344513,
-0.26702335476875305,
-0.07522575557231903,
0.19184096157550812,
0.16518841683864594,
0.2842872142791748,
-0.08664379268884659,
0.01121334359049797,
0.02237146906554699,
-0.06289319694042206,
-0.02038577012717724,
-0.05105176940560341,
0.10017417371273041,
-0.14920005202293396,
0.11657990515232086,
0.06752187758684158,
-0.020914411172270775,
0.03266291692852974,
-0.12070798128843307,
0.05268413946032524,
-0.0973779633641243,
0.07336171716451645,
0.1057887002825737,
-0.06971189379692078,
-0.020549653097987175,
0.13574528694152832,
-0.017145894467830658,
-0.2203558385372162,
-0.01696079783141613,
-0.039959296584129333,
-0.021819498389959335,
0.02137841284275055,
-0.09749490022659302,
0.033059123903512955,
0.12593132257461548,
0.00836427416652441,
0.09795163571834564,
0.13721811771392822,
-0.05460023507475853,
0.1458846628665924,
0.21675433218479156,
-0.007433003280311823,
0.018529340624809265,
-0.10358590632677078,
0.012855644337832928,
-0.03863892704248428,
0.0207180418074131,
-0.10036307573318481,
-0.0932718887925148,
0.013073299080133438,
0.052841175347566605,
-0.01090625487267971,
0.10885355621576309,
-0.031935714185237885,
0.08746455609798431,
0.06676541268825531,
-0.13182009756565094,
-0.1370803713798523,
0.04015682265162468,
-0.010030683130025864,
0.09730365127325058,
-0.015939880162477493,
0.08534841239452362,
-0.12495329976081848,
-0.02115805260837078,
-0.017359331250190735,
-0.058367662131786346,
-0.11155698448419571,
-0.08526475727558136,
0.07530451565980911,
0.05177298188209534,
-0.09251119941473007,
0.12922629714012146,
-0.0009436962427571416,
0.18169306218624115,
0.05396367236971855,
0.03944986313581467,
0.04369066655635834,
-0.05603845790028572,
0.08198993653059006,
0.14852482080459595,
-0.009778383187949657,
-0.04105912894010544,
-0.07215871661901474,
-0.10561059415340424,
0.08964990824460983,
-0.017719464376568794,
0.10432686656713486,
-0.10632530599832535,
-0.073843814432621,
0.007432502694427967,
0.031306639313697815,
-0.07393808662891388,
0.01965995877981186,
-0.028639988973736763,
-0.17428113520145416,
0.05944268777966499,
-0.0380004346370697,
0.06156166270375252,
-0.06680288165807724,
-0.11436987668275833,
-0.17618922889232635,
0.08190514892339706,
0.056641776114702225,
-0.051477402448654175,
-0.13643479347229004,
0.010903706774115562,
0.006213651038706303,
-0.03876681253314018,
-0.030653487890958786,
0.09431815892457962,
-0.11151596158742905,
0.042418185621500015,
-0.011852272786200047,
0.07420577108860016,
-0.06605134159326553,
-0.011078626848757267,
0.05766455829143524,
-0.09014534205198288,
-0.0017733711283653975,
0.033140916377305984,
-0.021665703505277634,
-0.016753679141402245,
-0.2402876317501068,
-0.02516130916774273,
0.029013313353061676,
0.010689162649214268,
0.09404846280813217,
-0.016692228615283966,
0.010369817726314068,
0.004810106009244919,
-0.09831332415342331,
0.01538196299225092,
0.14741139113903046,
-0.04067569226026535,
0.0052885026670992374,
0.05742860957980156,
-0.06302043050527573,
0.023809365928173065,
-0.023315787315368652,
0.11908557265996933,
0.03238719329237938,
0.06280376762151718,
-0.10495634377002716,
0.09348509460687637,
-0.12781332433223724,
-0.04360631853342056,
-0.02441735565662384,
0.07860706001520157,
0.037202138453722,
-0.10314428806304932,
0.008051802404224873,
0.03833739086985588,
0.19301730394363403,
-0.01708211936056614,
-0.011167561635375023,
-0.03700266778469086,
0.011281129904091358,
0.19605682790279388,
0.020971262827515602,
0.10650473088026047,
0.07163649797439575,
-0.0220173392444849,
0.13996753096580505,
0.1138046383857727,
0.11266620457172394,
-0.028778983280062675,
0.1301402598619461,
0.04734402522444725,
0.01927633211016655,
0.05886378884315491,
0.06016858294606209,
0.09117992967367172,
-0.010964158922433853,
0.07945526391267776,
0.028765598312020302,
-0.03266991674900055,
-0.037901654839515686,
0.1988201141357422,
0.11497367918491364,
-0.16505974531173706,
0.05737573653459549,
0.00018198277393821627,
0.03978715091943741,
-0.028985369950532913,
-0.1270243525505066,
-0.04439207538962364,
-0.31054916977882385,
0.1140095666050911,
-0.0630708560347557,
-0.005964191164821386,
0.040490612387657166,
-0.017254918813705444,
-0.002739418763667345,
-0.06700266897678375,
0.11672826111316681,
0.010379189625382423,
0.006502550560981035,
-0.03921150043606758,
-0.009089425206184387,
-0.10135465860366821,
-0.1012895330786705,
0.004339776933193207,
-0.16834335029125214,
-0.007342476863414049,
0.06098764017224312,
0.013856232166290283,
0.026889028027653694,
-0.029861140996217728,
-0.02155296318233013,
0.06141390651464462,
-0.03163966163992882,
0.050198208540678024,
0.12119366228580475,
0.10287556797266006,
-0.09907711297273636,
0.007870620116591454,
0.21069635450839996,
0.013911711983382702,
-0.14808879792690277,
-0.004182879813015461,
0.19718433916568756,
0.002850059187039733,
0.018603520467877388,
-0.02495824545621872,
-0.036406464874744415,
-0.06199747323989868,
0.1418970376253128,
0.10479418188333511,
-0.12811283767223358,
0.04452552646398544,
-0.07818030565977097,
-0.01041357684880495,
-0.06388362497091293,
0.11134158819913864,
0.11253110319375992,
0.04804401472210884,
-0.12215965986251831,
-0.10317116230726242,
-0.11707178503274918,
0.0028111140709370375,
-0.09418398141860962,
-0.0713438093662262,
0.06161827594041824,
-0.07260192930698395,
-0.07744508981704712,
0.02451702393591404,
-0.21036146581172943,
0.029039748013019562,
0.13437794148921967,
-0.23802228271961212,
-0.07594898343086243,
-0.08820874989032745,
0.11866119503974915,
0.017585763707756996,
0.09080028533935547,
-0.006763455457985401,
0.008966521359980106,
-0.1793689876794815,
-0.0017389452550560236,
-0.06428765505552292,
-0.025985464453697205,
0.08741623908281326,
-0.03367912396788597,
0.215544655919075,
-0.07904383540153503,
0.019456909969449043,
0.018524164333939552,
0.0882091075181961,
-0.040583495050668716,
0.0829564556479454,
-0.021467529237270355,
-0.10178262740373611,
-0.07720597833395004,
0.04307820275425911,
0.03392118215560913,
0.11213918030261993,
0.0703975185751915,
-0.07007893174886703,
0.034992218017578125,
0.10347025096416473,
0.030071692541241646,
-0.025086933746933937,
0.011146237142384052,
-0.13462688028812408,
0.04266480728983879,
-0.00950109213590622,
-0.06373808532953262,
0.02976018562912941,
-0.05616089329123497,
0.1345798671245575,
0.016780897974967957,
0.05263865366578102,
-0.07056242972612381,
0.02262748032808304,
-0.012564972043037415,
-0.05135739967226982,
-0.044162165373563766,
-0.21639741957187653,
-0.1230500265955925,
-0.131891131401062,
0.00044532367610372603,
-0.031320083886384964,
0.05634564161300659,
0.029048986732959747,
0.01923087239265442,
0.0021897810511291027,
-0.125748410820961,
0.035601187497377396,
0.12302692234516144,
-0.04143292456865311,
-0.038885973393917084
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "beomi/polyglot-ko-12.8b-safetensors"} | null | yatsby/koalpaca_persona_chat | [
"peft",
"arxiv:1910.09700",
"base_model:beomi/polyglot-ko-12.8b-safetensors",
"region:us"
] | 2024-02-06T07:34:40+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-beomi/polyglot-ko-12.8b-safetensors #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-beomi/polyglot-ko-12.8b-safetensors #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
39,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-beomi/polyglot-ko-12.8b-safetensors #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.10961789637804031,
0.21071495115756989,
-0.0027299211360514164,
0.030867598950862885,
0.08404036611318588,
0.023725878447294235,
0.05362171679735184,
0.12906980514526367,
-0.0072654057294130325,
0.10364974290132523,
0.07074826210737228,
0.1023382842540741,
0.11220181733369827,
0.22433406114578247,
0.004169307183474302,
-0.19532260298728943,
0.027144387364387512,
-0.08582145720720291,
-0.001951515907421708,
0.12579995393753052,
0.14198173582553864,
-0.09728114306926727,
0.08253170549869537,
-0.012125578708946705,
-0.00739028537645936,
-0.03895310312509537,
-0.07343468815088272,
-0.025311358273029327,
0.03894927725195885,
0.04084749147295952,
0.05668249353766441,
-0.012989448383450508,
0.09116912633180618,
-0.2677059769630432,
0.01824372261762619,
0.04681527614593506,
-0.002370678586885333,
0.0870867595076561,
0.10048118978738785,
-0.036451343446969986,
0.1259821355342865,
-0.034365374594926834,
0.13458138704299927,
0.08524630963802338,
-0.08727609366178513,
-0.2199966013431549,
-0.06682506948709488,
0.0871523916721344,
0.19276690483093262,
0.07039836794137955,
-0.03849121183156967,
0.13201378285884857,
-0.07764609903097153,
0.017779123038053513,
0.0364728681743145,
-0.08855634927749634,
-0.07144439220428467,
0.05904533341526985,
0.10448364168405533,
0.06500660628080368,
-0.12519952654838562,
-0.03388572484254837,
0.02762475423514843,
0.032563500106334686,
0.07009617984294891,
0.00837742816656828,
0.1601426750421524,
0.027743641287088394,
-0.14419560134410858,
-0.04561416432261467,
0.14399901032447815,
0.02392285130918026,
-0.037787433713674545,
-0.22885115444660187,
-0.006041498854756355,
-0.08251570910215378,
-0.03220018744468689,
-0.0479237399995327,
0.03619791939854622,
0.006660545710474253,
0.11302857100963593,
-0.035996198654174805,
-0.0884331688284874,
-0.017436113208532333,
0.09959235042333603,
0.04865099489688873,
0.023892207071185112,
-0.016529694199562073,
0.008905068971216679,
0.12715387344360352,
0.06953214108943939,
-0.12865103781223297,
-0.06382443755865097,
-0.08064403384923935,
-0.04439681023359299,
-0.03829577937722206,
0.047405458986759186,
0.052130669355392456,
0.05529081076383591,
0.2569456100463867,
-0.013336011208593845,
0.05906929075717926,
0.07977642118930817,
0.009970862418413162,
0.05052295699715614,
0.10406793653964996,
-0.05113248899579048,
-0.1736934334039688,
-0.015014939941465855,
0.09176675975322723,
-0.0024748879950493574,
-0.028906583786010742,
-0.05433245003223419,
0.03951209411025047,
0.03736186772584915,
0.10738500207662582,
0.10816911607980728,
-0.011643948033452034,
-0.08116048574447632,
-0.06601212918758392,
0.21086978912353516,
-0.14910359680652618,
0.05380825698375702,
0.022713210433721542,
-0.008864509873092175,
-0.047283560037612915,
0.009516754187643528,
0.018732581287622452,
-0.03295474871993065,
0.08493981510400772,
-0.06753210723400116,
-0.041501641273498535,
-0.12342754751443863,
-0.02109818160533905,
0.029766783118247986,
0.004450530745089054,
-0.03440415486693382,
-0.03449053317308426,
-0.0767604187130928,
-0.09292563050985336,
0.10794906318187714,
-0.06615377217531204,
-0.06019843369722366,
-0.025008654221892357,
-0.09068076312541962,
0.021632099524140358,
0.026688026264309883,
0.08175653964281082,
-0.03094218485057354,
0.04008888080716133,
-0.012208833359181881,
0.06427491456270218,
0.0755091980099678,
0.03310775384306908,
-0.06806680560112,
0.05931594967842102,
-0.19788667559623718,
0.08434105664491653,
-0.08033992350101471,
0.030711794272065163,
-0.16065369546413422,
-0.013814491219818592,
0.01715327613055706,
0.018907854333519936,
0.02825186774134636,
0.1630518138408661,
-0.20455129444599152,
-0.03382095322012901,
0.16080088913440704,
-0.09828726947307587,
-0.11514415591955185,
0.039423707872629166,
-0.044122468680143356,
0.16090714931488037,
0.020380724221467972,
-0.01120840199291706,
0.10574221611022949,
-0.15381361544132233,
-0.021329231560230255,
-0.017106806859374046,
0.00014001473027747124,
0.08197958022356033,
0.08665264397859573,
-0.0874040275812149,
0.02102486602962017,
0.015773169696331024,
-0.05512072145938873,
-0.023638121783733368,
-0.04173196852207184,
-0.10846956074237823,
0.007768488489091396,
-0.08065543323755264,
0.021786149591207504,
-0.008479880169034004,
-0.0902753546833992,
-0.004476609639823437,
-0.15804734826087952,
-0.04096972942352295,
0.08493475615978241,
0.009207706898450851,
-0.023684825748205185,
-0.1016947403550148,
0.05195791646838188,
-0.04400825873017311,
-0.02319393679499626,
-0.13666759431362152,
-0.0180270466953516,
0.020980941131711006,
-0.13062602281570435,
-0.005796021316200495,
-0.11193039268255234,
0.06830789148807526,
0.007369262166321278,
-0.05169573426246643,
-0.040381185710430145,
0.0010229938197880983,
0.00324498419649899,
-0.05640646442770958,
-0.23668277263641357,
-0.03458556905388832,
-0.046340253204107285,
0.16219553351402283,
-0.21920496225357056,
0.03653312474489212,
0.03577139973640442,
0.13000322878360748,
0.002955146599560976,
-0.06530460715293884,
0.02544480934739113,
-0.06849335134029388,
-0.02416531927883625,
-0.07575736194849014,
0.00019970831635873765,
-0.0037140469066798687,
-0.03497619181871414,
0.011572835966944695,
-0.10946114361286163,
-0.04679575189948082,
0.10077084600925446,
0.0717339813709259,
-0.14652717113494873,
0.006278356071561575,
-0.04017193242907524,
-0.06252461671829224,
-0.07650189101696014,
-0.06172649934887886,
0.10465316474437714,
0.05470710247755051,
0.043072428554296494,
-0.07283446937799454,
-0.07850677520036697,
0.009825624525547028,
-0.028584672138094902,
-0.01664198562502861,
0.11326209455728531,
0.0775851458311081,
-0.09472689777612686,
0.09679245948791504,
0.07302488386631012,
0.03440585359930992,
0.0959782525897026,
-0.025416195392608643,
-0.10466854274272919,
-0.02948915772140026,
0.0441046804189682,
0.010400284081697464,
0.17353816330432892,
-0.07298446446657181,
0.05142870545387268,
0.04543355107307434,
-0.03810090199112892,
0.046606775373220444,
-0.09795264154672623,
0.012654358521103859,
0.012474296614527702,
-0.015725847333669662,
0.024113992229104042,
-0.02616758458316326,
0.008332003839313984,
0.08086678385734558,
0.05501897260546684,
0.0288657508790493,
0.023592356592416763,
-0.036365095525979996,
-0.13427437841892242,
0.18485194444656372,
-0.09804315865039825,
-0.23246149718761444,
-0.15983572602272034,
0.05572933703660965,
0.04707939550280571,
-0.016289735212922096,
0.015833774581551552,
-0.05885312333703041,
-0.10621979832649231,
-0.08435294032096863,
-0.0010348419891670346,
0.03406079486012459,
-0.055412039160728455,
-0.06250058114528656,
0.04061831161379814,
0.0458969809114933,
-0.12390007078647614,
0.0329398475587368,
0.05735870450735092,
-0.01575865037739277,
0.003889166982844472,
0.05323357135057449,
0.08803986012935638,
0.1860392689704895,
-0.007011431269347668,
-0.0018090619705617428,
0.061153534799814224,
0.2751663029193878,
-0.15598337352275848,
0.11623071134090424,
0.12945710122585297,
-0.06596867740154266,
0.07558190077543259,
0.18511152267456055,
0.03007260523736477,
-0.09431684017181396,
0.02221504971385002,
0.028530197218060493,
-0.019348926842212677,
-0.2623227834701538,
-0.05779441446065903,
-0.016790667548775673,
-0.0817514955997467,
0.07675120234489441,
0.08768691122531891,
0.07666543126106262,
0.03591623902320862,
-0.0669713020324707,
-0.09160799533128738,
0.022096192464232445,
0.10592213273048401,
-0.028706077486276627,
0.008937512524425983,
0.08239905536174774,
-0.04036097228527069,
0.008568873628973961,
0.09794377535581589,
-0.0056713782250881195,
0.15557977557182312,
0.060123175382614136,
0.11541707813739777,
0.0744662657380104,
0.09284735471010208,
-0.003543650032952428,
0.030534252524375916,
0.014416548423469067,
0.025546669960021973,
0.017474308609962463,
-0.08105999231338501,
0.017321879044175148,
0.11666354537010193,
0.042523667216300964,
0.026300739496946335,
0.023527204990386963,
-0.04409561678767204,
0.04946175962686539,
0.19076022505760193,
0.020519789308309555,
-0.2040753960609436,
-0.07656841725111008,
0.062279343605041504,
-0.0845729187130928,
-0.14859908819198608,
-0.01613408327102661,
0.02855117805302143,
-0.1669328659772873,
0.01979210413992405,
-0.04620518907904625,
0.10286183655261993,
-0.09675747156143188,
-0.04640926048159599,
0.11420920491218567,
0.06366953998804092,
-0.01860068365931511,
0.0485813282430172,
-0.17575745284557343,
0.11644912511110306,
0.025454780086874962,
0.07361066341400146,
-0.09271698445081711,
0.1043776124715805,
-0.002036268590018153,
-0.01517715584486723,
0.1621897965669632,
0.00899295974522829,
-0.05399046465754509,
-0.08428946882486343,
-0.09985753893852234,
-0.008134081028401852,
0.09331066161394119,
-0.1324402242898941,
0.07868587970733643,
-0.02822175808250904,
-0.030056042596697807,
-0.004497057758271694,
-0.09174525737762451,
-0.12310896068811417,
-0.16983850300312042,
0.06211012601852417,
-0.09773717820644379,
0.028552230447530746,
-0.0871608704328537,
-0.057316526770591736,
0.0076920632272958755,
0.1779889166355133,
-0.22696411609649658,
-0.11000986397266388,
-0.15073513984680176,
-0.1113404855132103,
0.16247029602527618,
-0.03994130343198776,
0.08907848596572876,
-0.002718474483117461,
0.1626734584569931,
0.015553733333945274,
-0.01124879252165556,
0.09613403677940369,
-0.09226591140031815,
-0.1948133409023285,
-0.05481252446770668,
0.16287854313850403,
0.14243386685848236,
0.03331607207655907,
-0.009955585934221745,
0.027070241048932076,
-0.05832253023982048,
-0.12261415272951126,
0.02596508339047432,
0.1509297788143158,
0.07717521488666534,
-0.018656287342309952,
-0.01747133955359459,
-0.0974668487906456,
-0.06065184623003006,
-0.040390219539403915,
-0.007241677492856979,
0.19157086312770844,
-0.07410354912281036,
0.15938133001327515,
0.10064543038606644,
-0.05799967050552368,
-0.20672455430030823,
0.03386897221207619,
0.05149837210774422,
0.021419551223516464,
0.02438853681087494,
-0.197852224111557,
0.08761600404977798,
-0.012796605005860329,
-0.07696714997291565,
0.16622881591320038,
-0.1731831580400467,
-0.14170759916305542,
0.10386216640472412,
0.02707586996257305,
-0.20205776393413544,
-0.13851600885391235,
-0.09671349078416824,
-0.0169175136834383,
-0.1392814666032791,
0.05337141081690788,
0.012694485485553741,
0.008979608304798603,
0.01961846463382244,
0.022247226908802986,
0.023870712146162987,
-0.04658494517207146,
0.2135465145111084,
-0.025664545595645905,
0.0017215947154909372,
-0.04860874265432358,
-0.08420266211032867,
0.016618916764855385,
-0.05007964000105858,
0.10524231195449829,
0.0005444817361421883,
0.02580862306058407,
-0.17376582324504852,
-0.038285959511995316,
-0.05150719732046127,
0.026310471817851067,
-0.08997423946857452,
-0.08968009054660797,
-0.032617319375276566,
0.09162077307701111,
0.09386816620826721,
-0.02645791321992874,
0.0006103631458245218,
-0.08898759633302689,
0.05823022499680519,
0.19692958891391754,
0.19415283203125,
0.05697503685951233,
-0.05957503989338875,
0.017739567905664444,
-0.03079341910779476,
0.046935293823480606,
-0.22306200861930847,
0.04329647868871689,
0.05344805121421814,
0.021365638822317123,
0.0741555392742157,
-0.012966740876436234,
-0.15426652133464813,
-0.07309553027153015,
0.08345134556293488,
-0.060304194688797,
-0.16886289417743683,
-0.021161949262022972,
0.01897478476166725,
-0.209085151553154,
-0.03655313700437546,
0.027547743171453476,
-0.01919103041291237,
-0.04396763816475868,
0.02066725306212902,
0.07804011553525925,
-0.024425039067864418,
0.10184082388877869,
0.084506094455719,
0.09359288215637207,
-0.10419612377882004,
0.06966600567102432,
0.06729605048894882,
-0.02530461549758911,
0.029747575521469116,
0.10584407299757004,
-0.048107728362083435,
-0.031928908079862595,
0.07933232933282852,
0.09820890426635742,
0.021401626989245415,
-0.05679430067539215,
0.011688079684972763,
-0.05361153557896614,
0.05994763225317001,
0.1007290780544281,
0.031028956174850464,
0.0024281085934489965,
0.05738428980112076,
0.030467098578810692,
-0.09215402603149414,
0.11047801375389099,
0.06514456868171692,
0.01520591787993908,
-0.053895946592092514,
-0.0479254387319088,
-0.017461983487010002,
-0.015541561879217625,
-0.022235844284296036,
-0.00388848758302629,
-0.08485444635152817,
-0.0107139116153121,
-0.11201106756925583,
0.014453209936618805,
-0.07811205089092255,
0.011219043284654617,
0.02911180816590786,
-0.05066694691777229,
0.0022660864051431417,
0.001348224002867937,
-0.07413411885499954,
-0.05338117852807045,
-0.01667867600917816,
0.07831601798534393,
-0.12945392727851868,
0.04199930652976036,
0.07884632050991058,
-0.10563341528177261,
0.07440968602895737,
-0.003878686111420393,
0.012527674436569214,
0.0001688543998170644,
-0.14979718625545502,
0.05275307223200798,
-0.026627134531736374,
-0.010533341206610203,
0.019821448251605034,
-0.20480407774448395,
-0.0070218597538769245,
-0.041064441204071045,
-0.058160386979579926,
0.00761382794007659,
-0.010936712846159935,
-0.12206554412841797,
0.09925036877393723,
-0.004042772576212883,
-0.06322889029979706,
-0.0205230750143528,
0.04100993648171425,
0.09726599603891373,
-0.021457165479660034,
0.13416701555252075,
-0.022971071302890778,
0.07209911197423935,
-0.17231431603431702,
-0.007099600043147802,
-0.012736290693283081,
0.04353481903672218,
-0.02245442010462284,
-0.03326966241002083,
0.05931558832526207,
-0.01432125549763441,
0.18332786858081818,
-0.008577123284339905,
0.06727831810712814,
0.05135557800531387,
0.011832594871520996,
0.02528335154056549,
0.0799606665968895,
0.07013697922229767,
-0.0034178439527750015,
0.002012401120737195,
0.042154256254434586,
-0.00011239695595577359,
-0.041801679879426956,
-0.16172587871551514,
0.06145917624235153,
0.15863284468650818,
0.05912762135267258,
0.027486877515912056,
0.017242830246686935,
-0.12324082851409912,
-0.08110112696886063,
0.11303013563156128,
-0.03169916570186615,
-0.026108041405677795,
-0.07066348940134048,
0.172569140791893,
0.13314396142959595,
-0.2036903351545334,
0.06335624307394028,
-0.05075740069150925,
-0.04920513555407524,
-0.1407460868358612,
-0.1660064309835434,
-0.057182081043720245,
-0.05776215344667435,
-0.024081554263830185,
-0.06308441609144211,
0.048000846058130264,
0.04727386683225632,
0.0035547036677598953,
-0.017739716917276382,
0.10036874562501907,
0.017732392996549606,
-0.02941160276532173,
0.05497121065855026,
0.06511245667934418,
0.040414102375507355,
-0.08908241987228394,
0.006642209831625223,
-0.0042287432588636875,
0.013289268128573895,
0.07070746272802353,
0.01865154691040516,
-0.06310215592384338,
0.028609080240130424,
-0.013953626155853271,
-0.12718433141708374,
0.035968419164419174,
-0.013226272538304329,
-0.0373968780040741,
0.1561478078365326,
0.034185219556093216,
0.012566152028739452,
-0.021635454148054123,
0.2239960879087448,
-0.08157703280448914,
-0.06836149841547012,
-0.1497819423675537,
0.061931390315294266,
-0.07689252495765686,
0.030313808470964432,
0.03140432760119438,
-0.1209716945886612,
0.013141436502337456,
0.1688159555196762,
0.12624162435531616,
-0.016208605840802193,
0.008178540505468845,
0.046455543488264084,
0.005090147256851196,
-0.037005189806222916,
0.02670409344136715,
0.04723523184657097,
0.16760343313217163,
-0.07437742501497269,
0.05945286154747009,
-0.004837922751903534,
-0.08467333763837814,
-0.02004973217844963,
0.09626222401857376,
-0.003522216808050871,
-0.0018916050903499126,
-0.06500963866710663,
0.1379023939371109,
-0.08391589671373367,
-0.2110339105129242,
0.06814786046743393,
-0.0643806979060173,
-0.14525340497493744,
-0.03850167617201805,
0.03794271871447563,
-0.017018970102071762,
0.011048577725887299,
0.07238377630710602,
-0.047783929854631424,
0.20296555757522583,
0.037874117493629456,
-0.05310262367129326,
-0.08326008915901184,
0.05295826867222786,
-0.15470916032791138,
0.27560174465179443,
0.021786119788885117,
0.05045563727617264,
0.10891178250312805,
-0.020427634939551353,
-0.1494336873292923,
0.008533489890396595,
0.1067032441496849,
-0.06867165863513947,
0.05881962925195694,
0.16760657727718353,
0.004418961703777313,
0.1290789544582367,
0.06071395426988602,
-0.04438386857509613,
0.032971207052469254,
-0.09599718451499939,
-0.049756646156311035,
-0.11442965269088745,
0.08027500659227371,
-0.08588870614767075,
0.16035228967666626,
0.12455279380083084,
-0.0760989859700203,
-0.005720856133848429,
-0.017238695174455643,
0.08624733984470367,
0.007196058053523302,
0.1153140738606453,
0.006747027393430471,
-0.18312636017799377,
0.032771263271570206,
0.008125025779008865,
0.10596028715372086,
-0.1802573949098587,
-0.050434861332178116,
0.04364863410592079,
-0.019549986347556114,
-0.07929302752017975,
0.11982627213001251,
0.03376786410808563,
0.032387688755989075,
-0.03631647303700447,
-0.03690400347113609,
0.006775650195777416,
0.1413784772157669,
-0.11091645807027817,
-0.011496169492602348
] |
null | null | null |
# Lora of Dioscuri Pollux (Fate/Grand Order)
## What Is This?
This is the LoRA model of waifu Dioscuri Pollux (Fate/Grand Order).
## How Is It Trained?
* This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion).
* The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs).
* The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest).
* Dataset used for training is the `stage3-p480-800` in [CyberHarem/dioscuri_pollux_fgo](https://huggingface.co/datasets/CyberHarem/dioscuri_pollux_fgo), which contains 311 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 12, resolution is 720x720, clustering into 20 buckets.
* Trained for 3120 steps, 40 checkpoints were saved and evaluated.
* **Trigger word is `dioscuri_pollux_fgo`.**
* Pruned core tags for this waifu are `blonde_hair, bangs, breasts, medium_hair, blue_eyes, small_breasts`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
## How to Use It?
### If You Are Using A1111 WebUI v1.7+
**Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 2886, you need to download [`2886/dioscuri_pollux_fgo.pt`](https://huggingface.co/CyberHarem/dioscuri_pollux_fgo/resolve/main/2886/dioscuri_pollux_fgo.pt) as the embedding and [`2886/dioscuri_pollux_fgo.safetensors`](https://huggingface.co/CyberHarem/dioscuri_pollux_fgo/resolve/main/2886/dioscuri_pollux_fgo.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters.
## Which Step Should I Use?
We selected 5 good steps for you to choose. The best one is step 2886.
1600 images (1.62 GiB) were generated for auto-testing.

The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11).
Here are the preview of the recommended steps:
| Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0 | pattern_1 | pattern_2 | pattern_3_0 | pattern_3_1 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 |
|-------:|--------:|:----------|:-------------|:--------------|:----------|:------------------------------------------------------------------------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:----------------------------------------------|:----------------------------------------------|:--------------------------------------------|:--------------------------------------------|:--------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:--------------------------------|:------------------------------------|:--------------------------------|:----------------------------------|:----------------------------------------|:----------------------------------------|:----------------------------------------|:------------------------------|:----------------------------------|:----------------------------------|:--------------------------------|:------------------------------------------------|:----------------------------------|:----------------------------------|:------------------------------|:--------------------------------|:--------------------------------------|:--------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------|:--------------------------------------|
| 2886 | 38 | 0.883 | 0.856 | 0.814 | **0.770** | [Download](https://huggingface.co/CyberHarem/dioscuri_pollux_fgo/resolve/main/2886/dioscuri_pollux_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 1872 | 25 | **0.884** | 0.815 | 0.803 | 0.748 | [Download](https://huggingface.co/CyberHarem/dioscuri_pollux_fgo/resolve/main/1872/dioscuri_pollux_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 2496 | 33 | 0.846 | 0.816 | **0.814** | 0.718 | [Download](https://huggingface.co/CyberHarem/dioscuri_pollux_fgo/resolve/main/2496/dioscuri_pollux_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 3042 | 40 | 0.856 | **0.861** | 0.805 | 0.717 | [Download](https://huggingface.co/CyberHarem/dioscuri_pollux_fgo/resolve/main/3042/dioscuri_pollux_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
| 3120 | 41 | 0.838 | 0.832 | 0.810 | 0.701 | [Download](https://huggingface.co/CyberHarem/dioscuri_pollux_fgo/resolve/main/3120/dioscuri_pollux_fgo.zip) |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |  |
## Anything Else?
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
## All Steps
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* [Steps From 2418 to 3120](all/0.md)
* [Steps From 1638 to 2340](all/1.md)
* [Steps From 858 to 1560](all/2.md)
* [Steps From 78 to 780](all/3.md)
| {"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/dioscuri_pollux_fgo"], "pipeline_tag": "text-to-image"} | text-to-image | CyberHarem/dioscuri_pollux_fgo | [
"art",
"not-for-all-audiences",
"text-to-image",
"dataset:CyberHarem/dioscuri_pollux_fgo",
"license:mit",
"region:us"
] | 2024-02-06T07:36:38+00:00 | [] | [] | TAGS
#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/dioscuri_pollux_fgo #license-mit #region-us
| Lora of Dioscuri Pollux (Fate/Grand Order)
==========================================
What Is This?
-------------
This is the LoRA model of waifu Dioscuri Pollux (Fate/Grand Order).
How Is It Trained?
------------------
* This model is trained with HCP-Diffusion.
* The auto-training framework is maintained by DeepGHS Team.
* The base model used for training is deepghs/animefull-latest.
* Dataset used for training is the 'stage3-p480-800' in CyberHarem/dioscuri\_pollux\_fgo, which contains 311 images.
* Batch size is 4, resolution is 720x720, clustering into 5 buckets.
* Batch size for regularization dataset is 12, resolution is 720x720, clustering into 20 buckets.
* Trained for 3120 steps, 40 checkpoints were saved and evaluated.
* Trigger word is 'dioscuri\_pollux\_fgo'.
* Pruned core tags for this waifu are 'blonde\_hair, bangs, breasts, medium\_hair, blue\_eyes, small\_breasts'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable.
How to Use It?
--------------
### If You Are Using A1111 WebUI v1.7+
Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.
### If You Are Using A1111 WebUI v1.6 or Lower
After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.
For example, if you want to use the model from step 2886, you need to download '2886/dioscuri\_pollux\_fgo.pt' as the embedding and '2886/dioscuri\_pollux\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.
Which Step Should I Use?
------------------------
We selected 5 good steps for you to choose. The best one is step 2886.
1600 images (1.62 GiB) were generated for auto-testing.
!Metrics Plot
The base model used for generating preview images is Meina/MeinaMix\_V11.
Here are the preview of the recommended steps:
Anything Else?
--------------
Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:
1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.
2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.
3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.
4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.
5. Individuals who finds the generated image content offensive to their values.
All Steps
---------
We uploaded the files in all steps. you can check the images, metrics and download them in the following links:
* Steps From 2418 to 3120
* Steps From 1638 to 2340
* Steps From 858 to 1560
* Steps From 78 to 780
| [
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 2886, you need to download '2886/dioscuri\\_pollux\\_fgo.pt' as the embedding and '2886/dioscuri\\_pollux\\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 2886.\n\n\n1600 images (1.62 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 2418 to 3120\n* Steps From 1638 to 2340\n* Steps From 858 to 1560\n* Steps From 78 to 780"
] | [
"TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/dioscuri_pollux_fgo #license-mit #region-us \n",
"### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.",
"### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 2886, you need to download '2886/dioscuri\\_pollux\\_fgo.pt' as the embedding and '2886/dioscuri\\_pollux\\_fgo.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 2886.\n\n\n1600 images (1.62 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 2418 to 3120\n* Steps From 1638 to 2340\n* Steps From 858 to 1560\n* Steps From 78 to 780"
] | [
47,
38,
482
] | [
"passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/dioscuri_pollux_fgo #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file."
] | [
0.003411007346585393,
0.0033711898140609264,
-0.0044291079975664616,
0.08814991265535355,
0.07941135764122009,
0.06700671464204788,
0.21463388204574585,
0.08277364075183868,
0.1591861993074417,
-0.07257596403360367,
0.08370997756719589,
0.08130567520856857,
-0.014922045171260834,
0.05107172578573227,
-0.049759093672037125,
-0.16882683336734772,
-0.04539153724908829,
-0.01633213274180889,
-0.022244809195399284,
0.0064698453061282635,
0.08040063083171844,
0.004159328527748585,
0.10243391990661621,
-0.058332789689302444,
-0.037069182842969894,
0.04075442627072334,
-0.025120357051491737,
-0.03895841911435127,
0.02986769750714302,
0.09509726613759995,
0.10557869076728821,
0.03692742437124252,
0.06494299322366714,
-0.13517339527606964,
0.07019354403018951,
0.003906854894012213,
-0.12204837054014206,
0.004863554611802101,
0.012159289792180061,
-0.016482235863804817,
0.13356666266918182,
0.04308111593127251,
-0.09432276338338852,
0.037920668721199036,
-0.1367533802986145,
-0.05116216093301773,
-0.044429339468479156,
0.01591358333826065,
0.13298197090625763,
0.07451613992452621,
0.02166011556982994,
0.05725529417395592,
-0.06605658680200577,
0.09282591938972473,
0.10659987479448318,
-0.10474701970815659,
-0.06406363099813461,
0.10437356680631638,
0.02782868593931198,
0.1364733874797821,
-0.08313098549842834,
0.08000027388334274,
0.08298980444669724,
-0.04078257083892822,
-0.14712509512901306,
-0.09315814077854156,
-0.20669442415237427,
-0.02723631076514721,
-0.0021851162891834974,
0.0135654853656888,
0.42511504888534546,
0.0740799531340599,
0.027946753427386284,
0.06802225857973099,
-0.0796409323811531,
0.02300836518406868,
-0.09417499601840973,
0.12158672511577606,
0.04723126068711281,
0.10327084362506866,
-0.014704891480505466,
-0.11748695373535156,
-0.11404098570346832,
-0.07247405499219894,
-0.08873364329338074,
-0.019624071195721626,
0.015250180847942829,
0.11957304924726486,
-0.1968889832496643,
0.006097168195992708,
-0.06912858784198761,
-0.12973424792289734,
0.011620625853538513,
-0.10320992767810822,
0.17471708357334137,
0.05922408401966095,
-0.014173569157719612,
0.01974554732441902,
0.25505444407463074,
0.12312781810760498,
0.15960794687271118,
0.03135894238948822,
-0.11963947117328644,
0.123931884765625,
0.05808598920702934,
-0.08342757076025009,
-0.047925155609846115,
-0.0458095483481884,
0.16158059239387512,
-0.05784167721867561,
0.10047601908445358,
-0.05997309461236,
-0.11754769831895828,
0.023477187380194664,
-0.10245826840400696,
0.07057233154773712,
0.0532013364136219,
0.012565864250063896,
-0.05037648603320122,
0.04731662571430206,
0.044443197548389435,
-0.025554686784744263,
-0.0054212226532399654,
-0.008485265076160431,
-0.056524667888879776,
0.062127735465765,
0.11763973534107208,
0.03141626715660095,
0.059411585330963135,
0.006327627692371607,
-0.026677217334508896,
0.001062370603904128,
-0.041881173849105835,
-0.0023938468657433987,
0.05867399275302887,
0.05645633488893509,
0.08664073795080185,
-0.15454168617725372,
-0.079637810587883,
-0.016033221036195755,
0.034987881779670715,
-0.0014964024303480983,
0.08343014866113663,
-0.0028795686084777117,
0.04375206306576729,
-0.015146786347031593,
-0.021063873544335365,
0.02625895105302334,
-0.11077561229467392,
0.09822490811347961,
-0.033931925892829895,
0.09709136188030243,
-0.240634486079216,
-0.005045436322689056,
-0.05756997689604759,
0.01365672703832388,
0.021373102441430092,
0.004167009145021439,
-0.1040620356798172,
0.10459531098604202,
-0.002552652731537819,
0.0527590848505497,
-0.11032065004110336,
0.053830455988645554,
0.012420397251844406,
0.10029712319374084,
-0.10956887155771255,
0.0172292310744524,
0.12689104676246643,
-0.14821936190128326,
-0.14750885963439941,
0.0974738746881485,
-0.01774813048541546,
0.025431113317608833,
0.0466618649661541,
0.14356333017349243,
0.18141169846057892,
-0.20680765807628632,
-0.01862805150449276,
0.07033005356788635,
-0.03486914932727814,
-0.07701791822910309,
-0.014394029974937439,
0.10366109758615494,
0.026096699759364128,
0.03299969807267189,
-0.053934354335069656,
0.11800631135702133,
-0.036417536437511444,
-0.07399477809667587,
-0.03406009078025818,
-0.06935720890760422,
-0.07081438601016998,
0.05464660003781319,
-0.0218503400683403,
-0.0420977920293808,
0.025340348482131958,
-0.16342075169086456,
0.1537511944770813,
0.01147424802184105,
0.01906989887356758,
-0.0721580907702446,
0.14422042667865753,
0.011982052586972713,
0.007245290093123913,
-0.018342921510338783,
-0.08476511389017105,
-0.09795108437538147,
0.23009540140628815,
0.08299814909696579,
0.11419530212879181,
0.05626707524061203,
-0.048024605959653854,
-0.060864128172397614,
0.01786038652062416,
0.008942515589296818,
-0.04069855064153671,
0.02948639914393425,
-0.10901118814945221,
0.06135965883731842,
-0.024404587224125862,
0.041943639516830444,
0.0026962710544466972,
-0.02354544773697853,
0.1057368814945221,
0.02096341922879219,
-0.023503515869379044,
0.0855175107717514,
0.0394277460873127,
-0.02890062890946865,
-0.04897874593734741,
0.0014300403418019414,
0.07615606486797333,
-0.004112184047698975,
-0.06126688793301582,
0.0020429810974746943,
0.015034666284918785,
0.0720287635922432,
0.2117244005203247,
-0.20710515975952148,
0.028599584475159645,
0.019091453403234482,
0.036303263157606125,
0.054459117352962494,
-0.021649545058608055,
-0.045795634388923645,
0.03769408166408539,
-0.025926925241947174,
0.06302716583013535,
-0.027402816340327263,
0.06368616968393326,
-0.006709618493914604,
-0.133232980966568,
-0.019627654924988747,
-0.022683901712298393,
0.19527097046375275,
-0.15499147772789001,
0.07960701733827591,
0.17290066182613373,
-0.11585012823343277,
0.14311464130878448,
-0.01717373915016651,
-0.0030661055352538824,
0.01639651507139206,
0.05705108121037483,
-0.0046358900144696236,
0.10306957364082336,
-0.10634854435920715,
0.007010478060692549,
0.02183917723596096,
-0.08485527336597443,
0.03204220160841942,
-0.12223371118307114,
-0.11483769863843918,
-0.07383646816015244,
-0.046558596193790436,
-0.03881455957889557,
0.02824701927602291,
-0.03588496521115303,
0.09392819553613663,
-0.08919092267751694,
-0.09881792962551117,
-0.012266253121197224,
-0.07850571721792221,
0.019099872559309006,
0.023052837699651718,
-0.06389475613832474,
-0.12987010180950165,
-0.11394241452217102,
-0.05472148582339287,
-0.14919331669807434,
-0.0005554094677790999,
0.06989461183547974,
-0.10878996551036835,
-0.0477459616959095,
0.01651075668632984,
-0.050189901143312454,
0.10119887441396713,
-0.06947055459022522,
0.027637982740998268,
0.05232252553105354,
-0.034363843500614166,
-0.17008699476718903,
-0.012248314917087555,
-0.06995876878499985,
-0.06268933415412903,
0.14522644877433777,
-0.14527758955955505,
0.17722633481025696,
-0.030587023124098778,
0.052463918924331665,
0.06100961193442345,
0.04216315224766731,
0.17229074239730835,
-0.11489640176296234,
0.07652239501476288,
0.21016603708267212,
0.06379420310258865,
0.06907425075769424,
0.12776337563991547,
0.07758332043886185,
-0.12935183942317963,
0.035817503929138184,
0.07244906574487686,
-0.1011563241481781,
-0.08332335948944092,
-0.0711415708065033,
-0.10522487014532089,
-0.04227786511182785,
0.05274393782019615,
0.06170976161956787,
0.07790756970643997,
0.11898625642061234,
-0.06993184238672256,
0.010865193791687489,
0.11629131436347961,
0.0348578579723835,
0.09581810235977173,
0.0033985450863838196,
0.05144357308745384,
-0.13407671451568604,
-0.05974438041448593,
0.1637669801712036,
0.17990513145923615,
0.21756035089492798,
0.015990689396858215,
0.043298084288835526,
0.10355773568153381,
0.07486870884895325,
0.08245936781167984,
0.06728541105985641,
0.015236735343933105,
0.014933189377188683,
-0.07236027717590332,
-0.052662357687950134,
0.03480169549584389,
-0.00768905458971858,
-0.02349872514605522,
-0.15434607863426208,
0.08672494441270828,
-0.014989282004535198,
0.07645601034164429,
0.13994082808494568,
0.04130781069397926,
-0.11966182291507721,
0.15332801640033722,
0.08070918172597885,
0.10142531245946884,
-0.07337740063667297,
0.12300114333629608,
0.06556181609630585,
-0.009403577074408531,
0.1423708200454712,
0.027770111337304115,
0.13914990425109863,
-0.03610756993293762,
-0.06595483422279358,
-0.062454137951135635,
-0.04335683211684227,
0.008849174715578556,
0.02115173265337944,
-0.23156508803367615,
0.0958147794008255,
0.05650179460644722,
0.011796602047979832,
-0.0202227383852005,
-0.04135997220873833,
0.16655652225017548,
0.16917549073696136,
0.10426252335309982,
0.02116185426712036,
-0.0012686359696090221,
-0.02260817028582096,
-0.08323100209236145,
0.04363344609737396,
0.012692710384726524,
0.05000270530581474,
-0.04777521640062332,
-0.08372612297534943,
-0.022680850699543953,
0.005404343828558922,
0.024392513558268547,
-0.06428548693656921,
-0.10767467319965363,
-0.03863893076777458,
0.25259703397750854,
-0.022100461646914482,
0.04479563981294632,
0.05431352183222771,
0.0341687835752964,
-0.03960235416889191,
0.021258974447846413,
-0.0411500483751297,
-0.02778267301619053,
-0.008597842417657375,
0.003926227800548077,
0.005448098760098219,
-0.04410572722554207,
-0.06725291907787323,
-0.02179097570478916,
-0.10784193873405457,
-0.11090559512376785,
0.023466214537620544,
-0.05009065940976143,
-0.0027895013336092234,
-0.021216362714767456,
0.03752904385328293,
-0.09214036911725998,
-0.03536367788910866,
0.02827053889632225,
0.017939263954758644,
-0.04632499814033508,
-0.14257067441940308,
-0.01011438388377428,
-0.009576949290931225,
-0.026084527373313904,
0.035985130816698074,
-0.11181452125310898,
-0.07405432313680649,
-0.048202719539403915,
-0.014199494384229183,
0.11277998983860016,
0.22985197603702545,
-0.024194877594709396,
-0.015977999195456505,
0.1345074623823166,
-0.08584894984960556,
-0.3276314437389374,
-0.15050235390663147,
-0.1486552208662033,
-0.11757273972034454,
0.020450273528695107,
-0.06865518540143967,
0.028317445889115334,
0.10589494556188583,
-0.04257354885339737,
0.21495354175567627,
-0.1833612471818924,
-0.09391194581985474,
0.07904954999685287,
0.0977468192577362,
0.3098257780075073,
-0.23891954123973846,
0.015859657898545265,
-0.10452986508607864,
-0.04613139107823372,
0.051966190338134766,
-0.08903279900550842,
0.10186240077018738,
0.04972045496106148,
0.0494503453373909,
-0.001715097576379776,
-0.007568058092147112,
0.15144525468349457,
-0.08345843106508255,
0.14354459941387177,
-0.10763860493898392,
-0.09059539437294006,
0.19733767211437225,
-0.016588563099503517,
0.021554743871092796,
-0.16408850252628326,
-0.03879189491271973,
-0.009350311942398548,
0.03446807712316513,
-0.029557857662439346,
0.06482955068349838,
-0.009619554504752159,
-0.018748046830296516,
-0.1347372978925705,
-0.02068367972970009,
-0.042977068573236465,
0.05683469399809837,
0.21737268567085266,
-0.07538623362779617,
-0.05880076065659523,
0.06033177673816681,
0.021331511437892914,
0.10300379246473312,
0.008118964731693268,
-0.04532155022025108,
-0.04477592930197716,
0.09652145206928253,
-0.19042938947677612,
0.06406054645776749,
0.01355128176510334,
-0.010789922438561916,
0.025465916842222214,
0.024152396246790886,
0.01191608514636755,
0.10609106719493866,
0.16596758365631104,
-0.021045105531811714,
-0.05982779338955879,
-0.026858707889914513,
0.011250853538513184,
0.12311475723981857,
-0.012217194773256779,
0.10660555958747864,
0.022139975801110268,
0.03539609909057617,
0.0068557108752429485,
0.05378573760390282,
-0.07858027517795563,
-0.08400614559650421,
0.09679034352302551,
-0.047135498374700546,
-0.10544031113386154,
0.08070922642946243,
0.046254705637693405,
0.04642839729785919,
-0.006201261654496193,
0.03960094600915909,
0.010320354253053665,
-0.13186177611351013,
0.02026483230292797,
0.19666311144828796,
-0.0593002550303936,
-0.0626865029335022,
-0.0827937200665474,
-0.005584533791989088,
-0.11181148886680603,
0.11053839325904846,
0.031116722151637077,
-0.018602797761559486,
0.12281666696071625,
-0.04474223032593727,
-0.05663572624325752,
-0.0014733514981344342,
-0.07838602364063263,
0.02174706757068634,
-0.14045484364032745,
-0.1906113624572754,
0.03695826232433319,
-0.004364109598100185,
-0.06536997854709625,
-0.08369360864162445,
-0.0971023365855217,
0.054863110184669495,
-0.13247552514076233,
0.13435085117816925,
-0.06423021107912064,
0.05574523285031319,
-0.036669231951236725,
-0.05250006169080734,
-0.08814170956611633,
-0.016183221712708473,
-0.05940565466880798,
-0.0281949732452631,
0.058101192116737366,
0.005162718705832958,
-0.1304807811975479,
-0.1235409826040268,
0.04851514473557472,
-0.009055476635694504,
0.002648397348821163,
0.02456498332321644,
-0.06702008843421936,
0.011017532087862492,
-0.23627854883670807,
-0.05631544440984726,
0.10031599551439285,
0.04007034748792648,
-0.08024954050779343,
0.12940523028373718,
0.05088333785533905,
-0.013451234437525272,
0.027476688846945763,
0.008912624791264534,
0.14893478155136108,
-0.08019819855690002,
0.047377511858940125,
-0.12506479024887085,
-0.15721926093101501,
-0.02797030285000801,
0.027752233669161797,
0.24915331602096558,
0.09005098789930344,
0.12282858043909073,
-0.0581168606877327,
0.036060139536857605,
-0.03162289410829544,
0.06327372789382935,
0.021242434158921242,
-0.09884011000394821,
-0.029488440603017807,
-0.15591725707054138,
-0.05920650064945221,
-0.056190237402915955,
0.15106524527072906,
0.050870295614004135,
-0.10020512342453003,
0.0032424863893538713,
0.11067554354667664,
-0.1429983228445053,
-0.0021350309252738953,
0.14155179262161255,
-0.048399630934000015,
0.032539159059524536,
-0.14948832988739014,
0.045423462986946106,
0.09016850590705872,
-0.004136341158300638,
-0.0009998063324019313,
0.09393151104450226,
-0.010375616140663624,
0.00003559451943146996,
0.040419161319732666,
-0.0312812477350235,
0.10024072974920273,
-0.06467845290899277,
0.06883034855127335,
-0.006263361312448978,
-0.05584406480193138,
-0.14410832524299622,
0.15620963275432587,
-0.019920609891414642,
0.015960386022925377,
-0.04146866872906685,
-0.006786404177546501,
-0.11774972081184387,
-0.11749155074357986,
-0.06887771934270859,
-0.14517565071582794,
0.07243359088897705,
-0.07333143055438995,
0.014523247256875038,
-0.03244511038064957,
0.013054200448095798,
-0.09412714838981628,
0.004314052406698465,
-0.1909271478652954,
-0.055200591683387756,
0.02965582348406315,
-0.013440780341625214,
-0.018153784796595573,
-0.04309595003724098,
-0.03670152649283409,
0.027979861944913864,
-0.06914927810430527,
-0.06540581583976746,
0.06932437419891357,
0.08223892003297806,
0.057431578636169434,
-0.16566866636276245,
-0.10736344009637833,
-0.06876561790704727,
0.03646862134337425,
0.05341881141066551,
0.21005448698997498,
0.032975323498249054,
-0.016852691769599915,
0.060551341623067856,
0.13119791448116302,
0.01352385152131319,
-0.05234945937991142,
-0.0904889777302742,
-0.11257784813642502,
-0.13849300146102905,
-0.00011228778748773038,
-0.05669548735022545,
-0.03509759530425072,
0.014228587970137596,
0.23440919816493988,
0.16825422644615173,
-0.14288674294948578,
0.04690982401371002,
-0.0724843293428421,
0.035562317818403244,
-0.021049443632364273,
0.14468882977962494,
0.05940063297748566,
0.16523431241512299,
-0.0388185940682888,
-0.020373014733195305,
-0.05233275890350342,
0.024227233603596687,
-0.10844270139932632,
0.011200718581676483,
-0.01608113944530487,
-0.06919989734888077,
-0.05981571227312088,
0.1062636747956276,
-0.11770378798246384,
0.057286154478788376,
0.20148707926273346,
-0.14298352599143982,
-0.005260237492620945,
-0.04334854707121849,
-0.005957651883363724,
0.10679548233747482,
0.030439823865890503,
-0.08583559840917587,
-0.0273271631449461,
0.02616143226623535,
0.014027292840182781,
-0.18805216252803802,
-0.11218466609716415,
-0.015523112379014492,
-0.11122044920921326,
0.1321188509464264,
-0.0021378870587795973,
0.026339538395404816,
0.02983153611421585,
-0.06801089644432068,
-0.0076311444863677025,
0.1796267330646515,
0.0175081267952919,
-0.0522761307656765,
-0.03571153059601784,
-0.06751542538404465,
-0.10801000148057938,
0.06541354209184647,
0.07824257016181946,
0.013393952511250973,
-0.0012089896481484175,
0.14530909061431885,
-0.03451649099588394,
-0.04441937804222107,
0.14506296813488007,
-0.18818987905979156,
0.0821891725063324,
-0.0053633120842278,
-0.017482921481132507,
-0.07448694854974747,
-0.048755597323179245,
0.044323526322841644,
0.09206241369247437,
-0.1647883504629135,
-0.05065165087580681,
0.055300530046224594,
-0.09607590734958649,
0.06016364321112633,
0.040557198226451874,
-0.0708199292421341,
0.02186955325305462,
-0.12455099821090698,
-0.006591370329260826,
-0.1053856834769249,
0.025216981768608093,
0.17967981100082397,
-0.03847738355398178,
0.015286714769899845,
-0.10787330567836761,
0.04911656305193901,
-0.023243816569447517,
-0.036283910274505615,
-0.06257890909910202
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# speecht5_finetuned_voxpopuli_nl
This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the facebook/voxpopuli dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4559
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 4
- eval_batch_size: 2
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 4000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.5165 | 7.37 | 1000 | 0.4733 |
| 0.4942 | 14.73 | 2000 | 0.4604 |
| 0.4867 | 22.1 | 3000 | 0.4569 |
| 0.4882 | 29.47 | 4000 | 0.4559 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.2
- Datasets 2.14.7
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["facebook/voxpopuli"], "base_model": "microsoft/speecht5_tts", "model-index": [{"name": "speecht5_finetuned_voxpopuli_nl", "results": []}]} | text-to-audio | magus4450/speecht5_finetuned_voxpopuli_nl | [
"transformers",
"tensorboard",
"safetensors",
"speecht5",
"text-to-audio",
"generated_from_trainer",
"dataset:facebook/voxpopuli",
"base_model:microsoft/speecht5_tts",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-06T07:38:54+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us
| speecht5\_finetuned\_voxpopuli\_nl
==================================
This model is a fine-tuned version of microsoft/speecht5\_tts on the facebook/voxpopuli dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4559
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 4
* eval\_batch\_size: 2
* seed: 42
* gradient\_accumulation\_steps: 8
* total\_train\_batch\_size: 32
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* training\_steps: 4000
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.2
* Datasets 2.14.7
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 2\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.2\n* Datasets 2.14.7\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 2\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.2\n* Datasets 2.14.7\n* Tokenizers 0.15.0"
] | [
74,
158,
4,
32
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 2\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.2\n* Datasets 2.14.7\n* Tokenizers 0.15.0"
] | [
-0.13301466405391693,
0.06173160672187805,
-0.00251975329592824,
0.04919426888227463,
0.11387114226818085,
0.015134218148887157,
0.08025407046079636,
0.1373039186000824,
-0.08052506297826767,
0.09620442986488342,
0.07824365049600601,
0.05521128699183464,
0.06517021358013153,
0.18369431793689728,
-0.026869524270296097,
-0.3066260814666748,
-0.0008792130975052714,
-0.01211980264633894,
-0.14411289989948273,
0.11712735146284103,
0.0948990061879158,
-0.09957777708768845,
0.025589456781744957,
-0.017849106341600418,
-0.10797131061553955,
0.002195061184465885,
-0.014456472359597683,
-0.03876505047082901,
0.11398862302303314,
0.052948590368032455,
0.06916215270757675,
0.04101744666695595,
0.0936754122376442,
-0.2494356483221054,
0.021270180121064186,
0.08079349994659424,
0.02973056025803089,
0.06594311445951462,
0.11328110098838806,
-0.011822251603007317,
0.11845733225345612,
-0.0854206383228302,
0.0584777295589447,
0.04993771016597748,
-0.1102573499083519,
-0.32290294766426086,
-0.10122119635343552,
0.02634335495531559,
0.15467552840709686,
0.07922585308551788,
-0.04519983381032944,
0.05139598995447159,
-0.06539224833250046,
0.09752736240625381,
0.22972288727760315,
-0.22856555879116058,
-0.07711830735206604,
-0.00054655980784446,
0.10897359251976013,
0.08203910291194916,
-0.11495088785886765,
-0.0044968328438699245,
0.04027410224080086,
0.026446808129549026,
0.1081981509923935,
-0.006502619944512844,
0.04640743508934975,
-0.008292526006698608,
-0.14762942492961884,
-0.02673940919339657,
0.10415974259376526,
0.09192732721567154,
-0.024957576766610146,
-0.1160096675157547,
-0.018511051312088966,
-0.19413131475448608,
-0.04498732462525368,
0.012088148854672909,
0.02279084175825119,
-0.0407697968184948,
-0.10327915847301483,
-0.003202820895239711,
-0.06188466027379036,
-0.09337470680475235,
0.04502559453248978,
0.10180164873600006,
0.031853556632995605,
-0.046506669372320175,
0.02427455224096775,
0.11306790262460709,
0.02882150001823902,
-0.14156296849250793,
0.02266240492463112,
0.022094203159213066,
-0.12402815371751785,
-0.031429797410964966,
-0.028262851759791374,
-0.0477454774081707,
0.0018270604778081179,
0.14326123893260956,
-0.03782055154442787,
0.08672519028186798,
0.020517604425549507,
0.0336189828813076,
-0.06960377097129822,
0.11193899065256119,
-0.06569625437259674,
-0.11065898835659027,
-0.05348408594727516,
0.1129477322101593,
-0.008151598274707794,
-0.013075331225991249,
-0.06771598756313324,
0.022769808769226074,
0.06683463603258133,
0.05000236630439758,
-0.0011380480136722326,
0.007761165965348482,
-0.0901109054684639,
-0.02639235183596611,
0.004794894717633724,
-0.0889725461602211,
0.05404934287071228,
0.021283317357301712,
-0.04103211313486099,
-0.03943180665373802,
-0.0017289636889472604,
0.03330688551068306,
0.0031864934135228395,
0.1461828649044037,
-0.04229069501161575,
-0.014073836617171764,
-0.08477485179901123,
-0.10348931699991226,
0.037193771451711655,
-0.05182873085141182,
0.0037990340497344732,
-0.04467138648033142,
-0.10161837190389633,
-0.07532572746276855,
0.07142778486013412,
-0.04522173851728439,
-0.06718656420707703,
-0.05196511000394821,
-0.03835815191268921,
0.04518814757466316,
-0.039694663137197495,
0.19815732538700104,
-0.06947986036539078,
0.12088032811880112,
-0.004247321281582117,
0.0638805627822876,
0.05668674036860466,
0.08076919615268707,
-0.024744046851992607,
0.05794556066393852,
-0.19801512360572815,
0.07526151835918427,
-0.10550195723772049,
0.04217693954706192,
-0.13903380930423737,
-0.10529933869838715,
-0.024339457973837852,
0.022440098226070404,
0.08599395304918289,
0.09739844501018524,
-0.19286392629146576,
-0.11290464550256729,
0.16646355390548706,
-0.0761631578207016,
-0.0847131609916687,
0.14130914211273193,
-0.020069994032382965,
0.0015970611711964011,
0.03205404058098793,
0.1871209442615509,
0.11649259179830551,
-0.11604002863168716,
0.03228217363357544,
-0.042008887976408005,
0.1023278683423996,
0.04751117154955864,
0.09308827668428421,
-0.04324435442686081,
0.04058656841516495,
-0.022025350481271744,
-0.010087269358336926,
0.08099444955587387,
-0.07720788568258286,
-0.06449974328279495,
0.00027508524362929165,
-0.07567770034074783,
0.041153281927108765,
0.05755382776260376,
0.00397541094571352,
-0.09862618148326874,
-0.11889006942510605,
0.039730388671159744,
0.10094398260116577,
-0.07879139482975006,
0.032287776470184326,
-0.042331140488386154,
0.020820604637265205,
-0.0003105849027633667,
-0.01923234574496746,
-0.1681424379348755,
0.0027774113696068525,
0.01478639431297779,
-0.04820763319730759,
0.024036137387156487,
0.005433604586869478,
0.0716264620423317,
0.046583112329244614,
-0.09020925313234329,
-0.0712772086262703,
-0.04852674901485443,
0.009339947253465652,
-0.08463796228170395,
-0.2522042393684387,
-0.0662422776222229,
-0.03935513645410538,
0.1602751612663269,
-0.22377774119377136,
0.009145444259047508,
0.035254281014204025,
0.1431812196969986,
0.06257325410842896,
-0.04966957867145538,
0.019935626536607742,
0.08938513696193695,
0.0033350687008351088,
-0.0771542564034462,
0.034403666853904724,
0.005471378099173307,
-0.14353464543819427,
0.005009428597986698,
-0.12485092878341675,
0.09889169782400131,
0.08365748077630997,
0.0119871711358428,
-0.11304742842912674,
-0.08457406610250473,
-0.06423608958721161,
-0.0679028257727623,
-0.03588305041193962,
-0.0032308800145983696,
0.14569023251533508,
0.03347780182957649,
0.10503210872411728,
-0.07209000736474991,
-0.04368868097662926,
0.0364425852894783,
-0.00021457810362335294,
-0.006304427981376648,
0.15003804862499237,
0.038623590022325516,
-0.08910168707370758,
0.10869383066892624,
0.12548129260540009,
-0.033889930695295334,
0.18348492681980133,
-0.08058278262615204,
-0.11283145844936371,
-0.022993383929133415,
0.033697161823511124,
0.027365846559405327,
0.12279308587312698,
-0.11783825606107712,
0.010056449100375175,
0.012526795268058777,
0.035682715475559235,
0.011698133312165737,
-0.18715442717075348,
-0.017619194462895393,
0.04522239789366722,
-0.058690689504146576,
-0.022861026227474213,
-0.013329212553799152,
-0.003814413445070386,
0.0794963389635086,
0.015977833420038223,
-0.027604149654507637,
0.0006849412457086146,
-0.021992182359099388,
-0.08712141215801239,
0.16646891832351685,
-0.11284554749727249,
-0.15891923010349274,
-0.12535960972309113,
-0.036150012165308,
0.019115280359983444,
-0.018100520595908165,
0.06187381222844124,
-0.10381883382797241,
-0.021333245560526848,
-0.05174041911959648,
0.05124310031533241,
-0.0407119058072567,
0.020160840824246407,
-0.030672073364257812,
0.033771317452192307,
0.08253638446331024,
-0.07850667089223862,
0.042048078030347824,
-0.006244097836315632,
-0.010950841009616852,
0.011194175109267235,
0.02208801731467247,
0.07826725393533707,
0.15568435192108154,
0.05405423045158386,
0.0058981310576200485,
-0.058776725083589554,
0.14852622151374817,
-0.15453031659126282,
0.01966984011232853,
0.12963691353797913,
-0.015171397477388382,
0.042039260268211365,
0.16432517766952515,
0.047217294573783875,
-0.08100958168506622,
0.034461624920368195,
0.051364392042160034,
-0.018998512998223305,
-0.237263485789299,
-0.029892420396208763,
-0.06988269090652466,
0.0130030307918787,
0.09994097799062729,
0.029072029516100883,
-0.021913817152380943,
0.026333559304475784,
-0.031714290380477905,
-0.01124435756355524,
0.02471233904361725,
0.048757243901491165,
0.03342566266655922,
0.01683976873755455,
0.10667119175195694,
-0.020982414484024048,
-0.004470474552363157,
0.05637607350945473,
0.006624187808483839,
0.2384551763534546,
0.0012356676161289215,
0.14737334847450256,
0.058850377798080444,
0.13370844721794128,
0.010969417169690132,
0.034304697066545486,
0.017093297094106674,
-0.03140097111463547,
0.005484180990606546,
-0.051094770431518555,
-0.0011699242750182748,
0.04513882100582123,
0.07868268340826035,
0.0059679108671844006,
-0.1296156346797943,
-0.023442795500159264,
0.012838931754231453,
0.3152429163455963,
0.08915974199771881,
-0.2515547573566437,
-0.09300404042005539,
0.02186867780983448,
-0.0700928345322609,
-0.052187126129865646,
0.026444068178534508,
0.15124742686748505,
-0.09078068286180496,
0.08764423429965973,
-0.0792798176407814,
0.09480901062488556,
-0.05060873180627823,
0.0005796396289952099,
0.10325803607702255,
0.08027734607458115,
-0.02704201266169548,
0.05074001103639603,
-0.2518492639064789,
0.30445998907089233,
0.002260262379422784,
0.06710974872112274,
-0.01746310107409954,
0.03373222425580025,
0.03337491676211357,
-0.0056664664298295975,
0.10149910300970078,
-0.006890593096613884,
-0.16892755031585693,
-0.1716887354850769,
-0.07400189340114594,
-0.004359501414000988,
0.1347540318965912,
-0.06462621688842773,
0.08764500170946121,
-0.02748393826186657,
-0.025589678436517715,
0.047374505549669266,
-0.09159376472234726,
-0.10335154086351395,
-0.10797949135303497,
0.017588412389159203,
0.004785028751939535,
0.07146455347537994,
-0.11070588231086731,
-0.09853553026914597,
-0.048346903175115585,
0.16483868658542633,
-0.0814821645617485,
-0.012790332548320293,
-0.14965593814849854,
0.09985867142677307,
0.15499363839626312,
-0.053403954952955246,
0.07418633252382278,
0.02875356189906597,
0.107076495885849,
0.0068567777052521706,
-0.0006460649892687798,
0.14749592542648315,
-0.08178458362817764,
-0.21823137998580933,
-0.07946547865867615,
0.1921745091676712,
0.042890582233667374,
0.06935793906450272,
-0.03943922370672226,
0.033597953617572784,
0.0016695921076461673,
-0.06099877506494522,
0.0767512321472168,
-0.009079313836991787,
0.027572987601161003,
0.05899876728653908,
-0.032374490052461624,
-0.019630929455161095,
-0.029686331748962402,
-0.10110513120889664,
0.11229552328586578,
0.30860278010368347,
-0.08929399400949478,
0.06876026839017868,
0.06059420481324196,
-0.044257067143917084,
-0.17016860842704773,
0.050766367465257645,
0.13205106556415558,
0.05348043143749237,
0.06962840259075165,
-0.20572394132614136,
0.02456679753959179,
0.09102894365787506,
-0.024487387388944626,
0.08921761065721512,
-0.31461307406425476,
-0.13121210038661957,
0.07224372774362564,
0.0778970941901207,
-0.06370896100997925,
-0.14429914951324463,
-0.061537086963653564,
-0.0152204018086195,
-0.08572398871183395,
0.015598348341882229,
-0.07356639206409454,
0.13516756892204285,
0.021463701501488686,
0.012144590727984905,
0.02346695400774479,
-0.04372888803482056,
0.1252918690443039,
-0.02973092347383499,
0.06422111392021179,
-0.01227403711527586,
0.05220397189259529,
-0.029429765418171883,
-0.06094026193022728,
-0.03020801953971386,
-0.09665662795305252,
0.01138666644692421,
-0.11452851444482803,
-0.028287189081311226,
-0.07466577738523483,
0.029850680381059647,
-0.05296024680137634,
-0.049173079431056976,
-0.032493945211172104,
0.07409776002168655,
0.04860881343483925,
-0.01997794769704342,
0.11190589517354965,
-0.07786622643470764,
0.16783902049064636,
0.10476653277873993,
0.1038285568356514,
0.00967858824878931,
-0.10911603271961212,
-0.005634371656924486,
-0.031044628471136093,
0.04753011092543602,
-0.13413222134113312,
0.03711947426199913,
0.13385573029518127,
0.046937767416238785,
0.1462903469800949,
0.03847641125321388,
-0.0721091628074646,
0.024123776704072952,
0.08026587218046188,
-0.07331337034702301,
-0.1478227823972702,
-0.028047341853380203,
0.009233377873897552,
-0.13541476428508759,
-0.007958115078508854,
0.11071538925170898,
-0.020618868991732597,
-0.010197591967880726,
0.016359589993953705,
0.03113500028848648,
-0.047963764518499374,
0.22378219664096832,
0.02139550819993019,
0.08148311823606491,
-0.08894306421279907,
0.08323495090007782,
0.05030861496925354,
-0.18610359728336334,
0.022775594145059586,
0.1143590658903122,
-0.05560953542590141,
-0.010917257517576218,
0.05481249839067459,
0.07872691005468369,
0.05081885680556297,
-0.0335550494492054,
-0.09597095847129822,
-0.14172561466693878,
0.0683356449007988,
0.09979624301195145,
0.020357495173811913,
0.010698476806282997,
-0.015657372772693634,
0.04773637279868126,
-0.09268204122781754,
0.11725159734487534,
0.10764686018228531,
0.07545194029808044,
-0.14504307508468628,
0.1446036845445633,
0.007332217413932085,
-0.00949271209537983,
-0.012560846284031868,
0.019908154383301735,
-0.1157616376876831,
0.015384204685688019,
-0.10068108886480331,
-0.05508856475353241,
-0.0454886294901371,
-0.011302855797111988,
-0.010686259716749191,
-0.048290424048900604,
-0.03186914324760437,
0.016426416113972664,
-0.10455363243818283,
-0.0450921505689621,
-0.02181003801524639,
0.07001034915447235,
-0.08819127827882767,
-0.026000550016760826,
0.04038617014884949,
-0.09614931046962738,
0.08515069633722305,
0.01889856532216072,
0.03898333013057709,
0.00012588765821419656,
-0.11989935487508774,
0.001150667667388916,
0.028499482199549675,
-0.017024103552103043,
0.025748364627361298,
-0.16629858314990997,
-0.023959755897521973,
-0.05067262053489685,
0.026886871084570885,
-0.0038628578186035156,
-0.006080468185245991,
-0.1250499188899994,
-0.0003719837695825845,
-0.05264735966920853,
-0.06724320352077484,
-0.0503997728228569,
0.05054941028356552,
0.07142585515975952,
0.038388703018426895,
0.13937494158744812,
-0.09443620592355728,
0.05613565444946289,
-0.22708050906658173,
0.00293577928096056,
-0.021689005196094513,
-0.08011573553085327,
-0.05980789661407471,
-0.03568210080265999,
0.09272147715091705,
-0.049681514501571655,
0.07351912558078766,
-0.0374247282743454,
0.05724614858627319,
0.040544357150793076,
-0.1448954939842224,
0.056996606290340424,
0.0496848039329052,
0.1730605512857437,
0.021763937547802925,
-0.030277637764811516,
0.04691672697663307,
0.01660389080643654,
0.04671138525009155,
0.1538359373807907,
0.1517610251903534,
0.1731409877538681,
0.05312071368098259,
0.05803471431136131,
0.03924640640616417,
-0.1263876110315323,
-0.15487243235111237,
0.12143779546022415,
-0.023227576166391373,
0.14533279836177826,
-0.028745874762535095,
0.20241901278495789,
0.07924368977546692,
-0.214698925614357,
0.06361458450555801,
-0.04625300318002701,
-0.08267921954393387,
-0.08948659896850586,
-0.037435710430145264,
-0.07483198493719101,
-0.19785591959953308,
-0.0004266116302460432,
-0.0931987538933754,
0.05852743610739708,
0.03014255128800869,
0.03398514166474342,
0.03748684376478195,
0.1417696475982666,
0.028617732226848602,
-0.000460322160506621,
0.12139648199081421,
0.027377542108297348,
-0.0039973813109099865,
-0.037868399173021317,
-0.09710725396871567,
0.059242233633995056,
-0.05544018745422363,
0.029996419325470924,
-0.06175083667039871,
-0.11936269700527191,
0.06719780713319778,
0.034013450145721436,
-0.1073586717247963,
0.024532770738005638,
0.001406909548677504,
0.06859301775693893,
0.10848061740398407,
0.025468720123171806,
0.0024628133978694677,
-0.027259524911642075,
0.2561068832874298,
-0.10195515304803848,
-0.037960730493068695,
-0.12879039347171783,
0.22226029634475708,
-0.02414441667497158,
-0.00514176907017827,
0.004507587756961584,
-0.08182679861783981,
0.008322355337440968,
0.14319530129432678,
0.14102719724178314,
-0.02871246077120304,
-0.018788142129778862,
0.024716230109333992,
-0.010466717183589935,
-0.04611106216907501,
0.08039397746324539,
0.10957168787717819,
0.05589396879076958,
-0.05746348202228546,
-0.028530320152640343,
-0.02761031500995159,
-0.0671858862042427,
-0.01933852583169937,
0.07152505218982697,
0.04845697060227394,
0.0011345853563398123,
-0.012302596122026443,
0.15104447305202484,
-0.024374615401029587,
-0.16165316104888916,
0.03377620130777359,
-0.178136944770813,
-0.18332403898239136,
-0.04982883110642433,
0.08756877481937408,
0.036150481551885605,
0.04262080043554306,
0.0035285453777760267,
-0.012897377833724022,
0.0922926515340805,
-0.004011411219835281,
-0.013586252927780151,
-0.13206903636455536,
0.07385244965553284,
-0.10202106833457947,
0.17256614565849304,
-0.05260176956653595,
0.004821811802685261,
0.10588042438030243,
0.04608885571360588,
-0.06613089144229889,
0.03120596893131733,
0.07650229334831238,
-0.13556697964668274,
0.039745498448610306,
0.1994929313659668,
-0.05302273482084274,
0.15024186670780182,
0.04446709156036377,
-0.12707842886447906,
0.040903490036726,
-0.12117549777030945,
-0.08404436707496643,
-0.05652204155921936,
0.015659833326935768,
-0.04102150723338127,
0.1459846794605255,
0.20919659733772278,
-0.06178170070052147,
-0.007280530873686075,
-0.05306604504585266,
0.00577550707384944,
0.06275368481874466,
0.16970734298229218,
-0.010674817487597466,
-0.2577948272228241,
0.020236333832144737,
0.06495904922485352,
0.012842332012951374,
-0.23080195486545563,
-0.0977642610669136,
0.00860330369323492,
-0.04748889058828354,
-0.07003092020750046,
0.11939667910337448,
0.04914378747344017,
0.03664224222302437,
-0.04965359717607498,
-0.15299707651138306,
-0.02349078841507435,
0.1745450645685196,
-0.16501951217651367,
-0.04573764652013779
] |
null | null | null |
<!-- header start -->
<!-- 200823 -->
<div style="width: auto; margin-left: auto; margin-right: auto">
<img src="https://github.com/second-state/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
</div>
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
<!-- header end -->
# Qwen1.5-0.5B-Chat-GGUF
## Original Model
[Qwen/Qwen1.5-0.5B-Chat](https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat)
## Run with LlamaEdge
- LlamaEdge version: [v0.2.15](https://github.com/second-state/LlamaEdge/releases/tag/0.2.15) and above
- Prompt template
- Prompt type: `chatml`
- Prompt string
```text
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
```
- Run as LlamaEdge service
```bash
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Qwen1.5-0.5B-Chat-Q5_K_M.gguf llama-api-server.wasm -p chatml
```
- Run as LlamaEdge command app
```bash
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Qwen1.5-0.5B-Chat-Q5_K_M.gguf llama-chat.wasm -p chatml
```
## Quantized GGUF Models
| Name | Quant method | Bits | Size | Use case |
| ---- | ---- | ---- | ---- | ----- |
| [Qwen1.5-0.5B-Chat-Q2_K.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q2_K.gguf) | Q2_K | 2 | 298 MB| smallest, significant quality loss - not recommended for most purposes |
| [Qwen1.5-0.5B-Chat-Q3_K_L.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q3_K_L.gguf) | Q3_K_L | 3 | 364 MB| small, substantial quality loss |
| [Qwen1.5-0.5B-Chat-Q3_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q3_K_M.gguf) | Q3_K_M | 3 | 350 MB| very small, high quality loss |
| [Qwen1.5-0.5B-Chat-Q3_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q3_K_S.gguf) | Q3_K_S | 3 | 333 MB| very small, high quality loss |
| [Qwen1.5-0.5B-Chat-Q4_0.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q4_0.gguf) | Q4_0 | 4 | 395 MB| legacy; small, very high quality loss - prefer using Q3_K_M |
| [Qwen1.5-0.5B-Chat-Q4_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q4_K_M.gguf) | Q4_K_M | 4 | 407 MB| medium, balanced quality - recommended |
| [Qwen1.5-0.5B-Chat-Q4_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q4_K_S.gguf) | Q4_K_S | 4 | 397 MB| small, greater quality loss |
| [Qwen1.5-0.5B-Chat-Q5_0.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q5_0.gguf) | Q5_0 | 5 | 453 MB| legacy; medium, balanced quality - prefer using Q4_K_M |
| [Qwen1.5-0.5B-Chat-Q5_K_M.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q5_K_M.gguf) | Q5_K_M | 5 | 459 MB| large, very low quality loss - recommended |
| [Qwen1.5-0.5B-Chat-Q5_K_S.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q5_K_S.gguf) | Q5_K_S | 5 | 453 MB| large, low quality loss - recommended |
| [Qwen1.5-0.5B-Chat-Q6_K.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q6_K.gguf) | Q6_K | 6 | 515 MB| very large, extremely low quality loss |
| [Qwen1.5-0.5B-Chat-Q8_0.gguf](https://huggingface.co/second-state/Qwen1.5-0.5B-Chat-GGUF/blob/main/Qwen1.5-0.5B-Chat-Q8_0.gguf) | Q8_0 | 8 | 665 MB| very large, extremely low quality loss - not recommended |
| {"language": ["en"], "license": "other", "tags": ["chat"], "model_name": "Qwen1.5 0.5B Chat", "base_model": "Qwen/Qwen1.5-0.5B-Chat", "license_name": "tongyi-qianwen-research", "license_link": "https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat/blob/main/LICENSE", "model_creator": "Qwen", "quantized_by": "Second State Inc.", "pipeline_tag": "text-generation"} | text-generation | second-state/Qwen1.5-0.5B-Chat-GGUF | [
"gguf",
"chat",
"text-generation",
"en",
"base_model:Qwen/Qwen1.5-0.5B-Chat",
"license:other",
"region:us"
] | 2024-02-06T07:38:55+00:00 | [] | [
"en"
] | TAGS
#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-0.5B-Chat #license-other #region-us
|

---
Qwen1.5-0.5B-Chat-GGUF
======================
Original Model
--------------
Qwen/Qwen1.5-0.5B-Chat
Run with LlamaEdge
------------------
* LlamaEdge version: v0.2.15 and above
* Prompt template
+ Prompt type: 'chatml'
+ Prompt string
* Run as LlamaEdge service
* Run as LlamaEdge command app
Quantized GGUF Models
---------------------
| [] | [
"TAGS\n#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-0.5B-Chat #license-other #region-us \n"
] | [
39
] | [
"passage: TAGS\n#gguf #chat #text-generation #en #base_model-Qwen/Qwen1.5-0.5B-Chat #license-other #region-us \n"
] | [
-0.012140069156885147,
-0.008528387174010277,
-0.003774038515985012,
-0.014454744756221771,
0.013880643993616104,
-0.0008524732547812164,
0.1867283433675766,
0.08693931996822357,
0.19908417761325836,
-0.05094520002603531,
0.1026965081691742,
-0.040818627923727036,
0.05889695882797241,
0.11655580997467041,
0.03288577124476433,
-0.17059744894504547,
0.06700202822685242,
-0.04517889395356178,
-0.00795056577771902,
0.05798295885324478,
0.06590023636817932,
0.017564017325639725,
0.09401915967464447,
-0.0033853945787996054,
-0.09737144410610199,
0.029546435922384262,
-0.0030345702543854713,
-0.03479664400219917,
0.0836542621254921,
0.10340490937232971,
-0.01912570558488369,
0.10471117496490479,
-0.05418730154633522,
-0.22869718074798584,
0.040950972586870193,
-0.043552201241254807,
-0.13841399550437927,
-0.005745479837059975,
0.0029233423992991447,
0.023637738078832626,
0.0822356715798378,
0.1867697685956955,
-0.08954627066850662,
0.13088594377040863,
-0.13022755086421967,
-0.06773607432842255,
-0.07802162319421768,
0.09522989392280579,
0.011133870109915733,
0.05365408957004547,
-0.03158104047179222,
0.05193141847848892,
-0.12306715548038483,
0.022122196853160858,
0.10791710019111633,
-0.48932141065597534,
0.05387739837169647,
0.27870747447013855,
0.11071997135877609,
0.13641761243343353,
-0.08003716915845871,
0.04903245344758034,
0.05757977068424225,
-0.012609556317329407,
-0.17277979850769043,
-0.018573958426713943,
0.07061147689819336,
0.06353381276130676,
-0.04719134420156479,
-0.009384536184370518,
0.252652645111084,
0.0823051854968071,
-0.010150971822440624,
0.019043009728193283,
-0.012724525295197964,
-0.027110014110803604,
-0.026114635169506073,
0.06093672662973404,
-0.04331173747777939,
0.17767071723937988,
-0.0024421685375273228,
-0.08233266323804855,
-0.10472483932971954,
-0.055009447038173676,
-0.18052667379379272,
0.012498822063207626,
0.007604546844959259,
0.09318429231643677,
-0.1775340586900711,
0.0543302483856678,
-0.1679118573665619,
-0.07811977714300156,
-0.08321385085582733,
-0.1339317411184311,
0.029951948672533035,
0.0743471309542656,
0.001749400864355266,
0.12057333439588547,
0.2136716991662979,
0.14453306794166565,
-0.08460178971290588,
0.07178904116153717,
-0.0637650266289711,
0.10067646205425262,
0.029172776266932487,
-0.0039998795837163925,
0.05245888605713844,
-0.02550210990011692,
0.05219866335391998,
-0.09142526239156723,
-0.022745829075574875,
-0.0448782742023468,
-0.16511936485767365,
-0.0013240576954558492,
-0.11975142359733582,
0.08226680010557175,
0.007905459962785244,
0.023144476115703583,
0.006485832389444113,
-0.001308897859416902,
0.2190903276205063,
0.04764366149902344,
-0.06356897950172424,
0.05693120136857033,
-0.006725373677909374,
-0.08602089434862137,
-0.015244281850755215,
0.09583034366369247,
0.059573426842689514,
-0.13302771747112274,
-0.057529233396053314,
-0.09130801260471344,
0.037493038922548294,
0.032222870737314224,
0.02730191871523857,
0.04989228770136833,
0.021607592701911926,
-0.14754925668239594,
-0.08057811111211777,
0.04042236879467964,
-0.0011174401734024286,
-0.03234754875302315,
-0.08450248837471008,
-0.05537857115268707,
-0.051297564059495926,
0.00010343623580411077,
-0.059800054877996445,
-0.00016724127635825425,
-0.09376060962677002,
0.07698866724967957,
-0.04622440040111542,
0.007810864131897688,
-0.2560079097747803,
0.004280972760170698,
-0.03283974528312683,
0.020930005237460136,
-0.06926871091127396,
0.10557349026203156,
-0.17713546752929688,
0.08938848227262497,
-0.031017912551760674,
0.04532342031598091,
-0.09194081276655197,
0.008652061223983765,
-0.0684886947274208,
0.16630196571350098,
-0.11052551120519638,
-0.09506405144929886,
0.2841748595237732,
-0.09650584310293198,
-0.12644723057746887,
0.091712087392807,
-0.0026539317332208157,
-0.030074821785092354,
0.05176880210638046,
0.384842187166214,
0.05066786706447601,
-0.06936731189489365,
-0.030112827196717262,
0.1797925978899002,
-0.10570753365755081,
-0.06951157748699188,
0.08612152934074402,
-0.08203887194395065,
-0.09404883533716202,
0.04380635917186737,
0.011050748638808727,
0.13716357946395874,
-0.05286652594804764,
-0.06897204369306564,
-0.03642313554883003,
-0.059847909957170486,
0.02662607654929161,
-0.03772793710231781,
0.06772972643375397,
-0.056856315582990646,
-0.021573878824710846,
-0.11657267063856125,
0.08196263760328293,
0.08253472298383713,
-0.024378987029194832,
-0.13510861992835999,
0.06861771643161774,
0.09761172533035278,
0.05259005352854729,
-0.003933243453502655,
-0.09081367403268814,
0.019967859610915184,
0.049207959324121475,
0.09558027982711792,
0.11638087034225464,
0.056135743856430054,
-0.03327914699912071,
-0.00014779664343222976,
-0.006129324436187744,
0.025171369314193726,
-0.01057005487382412,
-0.03750865161418915,
-0.10738618671894073,
0.034211333841085434,
-0.04119192063808441,
0.11262239515781403,
-0.08743584156036377,
-0.0008142493315972388,
0.10252071917057037,
0.046357620507478714,
-0.0542110837996006,
0.03424164652824402,
-0.015385255217552185,
-0.040644578635692596,
0.003155442187562585,
-0.0028374367393553257,
0.07081636786460876,
0.03595512732863426,
-0.131589874625206,
0.22637920081615448,
-0.042573679238557816,
0.08730316907167435,
0.12584668397903442,
-0.009152332320809364,
0.07582326233386993,
-0.11832941323518753,
-0.0625734031200409,
-0.0069297426380217075,
0.10485925525426865,
-0.015133672393858433,
0.22375278174877167,
-0.04495280981063843,
0.07821758836507797,
-0.07185856252908707,
-0.04330158233642578,
-0.01323266327381134,
0.010791662149131298,
-0.023001927882432938,
0.10130183398723602,
0.09399551898241043,
-0.18086577951908112,
0.097293920814991,
0.17130370438098907,
0.22666531801223755,
0.26880916953086853,
-0.08893616497516632,
0.0451044887304306,
-0.01098586618900299,
0.061158619821071625,
-0.08470219373703003,
0.12047100067138672,
-0.21108755469322205,
-0.01259860210120678,
0.030215729027986526,
0.05780318006873131,
0.1305360645055771,
-0.05789598822593689,
-0.11272364854812622,
-0.05148701369762421,
-0.10778524726629257,
-0.14202333986759186,
0.1074645146727562,
-0.03492411598563194,
0.05668046325445175,
0.02385834977030754,
-0.019286615774035454,
0.07837920635938644,
-0.03651529550552368,
-0.07795796543359756,
0.1351599097251892,
-0.1517271101474762,
-0.15834298729896545,
-0.034168776124715805,
-0.05647728592157364,
-0.09666337817907333,
-0.004564824514091015,
0.10869172215461731,
-0.1775476336479187,
0.009612076915800571,
0.006352518685162067,
-0.02645854279398918,
-0.09998912364244461,
0.0126072708517313,
0.013438185676932335,
-0.016072530299425125,
-0.040373317897319794,
-0.1565592736005783,
-0.037027958780527115,
-0.03595005348324776,
-0.1190129891037941,
0.08429352939128876,
-0.16510140895843506,
0.011957257054746151,
0.1298549622297287,
0.0543598011136055,
0.09066513925790787,
-0.052468542009592056,
0.3540291488170624,
-0.05004870519042015,
-0.009503639303147793,
0.06726562231779099,
0.09318295121192932,
0.054493725299835205,
0.12928789854049683,
0.0230263639241457,
-0.0959409549832344,
0.0022904544603079557,
0.01797630824148655,
-0.06633191555738449,
-0.18205972015857697,
-0.10989338904619217,
-0.030042987316846848,
0.152581587433815,
-0.09771523624658585,
0.14939925074577332,
0.10693518072366714,
-0.021268298849463463,
0.008818695321679115,
-0.06962914019823074,
0.05779586359858513,
0.03542417660355568,
0.03245890885591507,
-0.04798170179128647,
0.06617596745491028,
-0.06363757699728012,
0.0768764466047287,
0.14291486144065857,
0.10943932831287384,
0.11517642438411713,
0.13980510830879211,
0.07452814280986786,
0.12101545184850693,
0.27478843927383423,
0.19310255348682404,
-0.15216349065303802,
-0.012968631461262703,
-0.0829073116183281,
-0.03234758600592613,
-0.027575820684432983,
0.02216130867600441,
0.007907967083156109,
-0.0012452186783775687,
-0.1927507519721985,
-0.040409136563539505,
-0.06821069121360779,
0.043757569044828415,
0.01946537382900715,
0.03392332047224045,
0.08941145986318588,
-0.007338224444538355,
0.043571244925260544,
-0.025447875261306763,
-0.001716612372547388,
0.13865746557712555,
0.07690492272377014,
-0.10538198798894882,
0.02525889314711094,
0.030812235549092293,
0.1039239764213562,
0.06764117628335953,
0.04825461283326149,
-0.07957010716199875,
-0.0689820647239685,
0.0046840510331094265,
0.10957629978656769,
-0.2096671611070633,
0.21192440390586853,
0.0030975034460425377,
-0.03681948408484459,
-0.03388475999236107,
-0.08080819994211197,
0.011884584091603756,
0.1254344880580902,
0.09431321173906326,
0.07067137956619263,
-0.09390047937631607,
0.027812840417027473,
-0.06257428228855133,
0.10053237527608871,
0.0755341425538063,
-0.030695267021656036,
-0.08825239539146423,
-0.003476993180811405,
0.029136022552847862,
-0.0025995452888309956,
0.052261535078287125,
-0.1596643030643463,
-0.09705722332000732,
-0.022281313315033913,
0.09070423245429993,
0.10334672778844833,
-0.0591018907725811,
0.06315239518880844,
-0.07786188274621964,
0.08514705300331116,
-0.102432020008564,
-0.06925597041845322,
-0.044274695217609406,
-0.07079756259918213,
0.013354551047086716,
-0.05240900442004204,
-0.01329125091433525,
-0.06052863225340843,
-0.01637752167880535,
-0.09719996899366379,
-0.22068746387958527,
0.06339952349662781,
-0.08072786778211594,
-0.05402793362736702,
-0.007054086308926344,
0.1995868682861328,
0.04310839995741844,
0.006115575321018696,
0.04345780983567238,
-0.09731297194957733,
-0.07118574529886246,
-0.19427171349525452,
0.04568561539053917,
-0.10134424269199371,
-0.11903002858161926,
0.007481027394533157,
0.017621885985136032,
-0.01597900502383709,
-0.031076965853571892,
-0.06866978853940964,
0.16705921292304993,
0.3129429817199707,
-0.06372992694377899,
0.21909712255001068,
0.2632904350757599,
-0.05794784799218178,
-0.13800612092018127,
-0.19340457022190094,
-0.09180726855993271,
-0.08154294639825821,
0.019165966659784317,
-0.16902731359004974,
-0.01644034869968891,
0.061364129185676575,
-0.052090488374233246,
0.14065033197402954,
-0.2901197373867035,
-0.04175791144371033,
0.09141827374696732,
-0.07690098136663437,
0.4291088283061981,
-0.1362432837486267,
-0.17410492897033691,
0.004969938658177853,
-0.25939053297042847,
0.14923015236854553,
-0.13369183242321014,
0.0955718383193016,
0.014267899096012115,
0.09054234623908997,
-0.03769320249557495,
0.013719527050852776,
0.13557010889053345,
0.04380054399371147,
0.008029380813241005,
-0.0822460725903511,
-0.057087525725364685,
0.06551714986562729,
0.040676165372133255,
-0.03365349397063255,
-0.18384051322937012,
0.056095853447914124,
-0.14385749399662018,
0.013555585406720638,
-0.12447318434715271,
0.028934309259057045,
-0.007923341356217861,
-0.05958056449890137,
-0.09932117909193039,
0.023643849417567253,
-0.04324601590633392,
0.031174365431070328,
0.108692966401577,
-0.08475179970264435,
0.12578190863132477,
0.11461614817380905,
-0.033865272998809814,
-0.1936902403831482,
-0.008228093385696411,
-0.10130436718463898,
-0.06116826459765434,
0.07481889426708221,
-0.19480295479297638,
-0.031334783881902695,
0.09035037457942963,
-0.03542553260922432,
0.07729462534189224,
0.050684213638305664,
-0.10645683109760284,
0.11916211247444153,
0.1203935295343399,
-0.14327508211135864,
-0.22490310668945312,
-0.06530563533306122,
0.10235261172056198,
0.16588377952575684,
0.07110749930143356,
0.10061642527580261,
0.024576954543590546,
-0.019310304895043373,
0.020148586481809616,
-0.004349615890532732,
-0.11584222316741943,
-0.04858482629060745,
0.047257665544748306,
0.008716210722923279,
-0.1557122766971588,
0.11597197502851486,
0.01393128838390112,
0.10446364432573318,
0.021357743069529533,
0.153960183262825,
-0.14505614340305328,
-0.08352165669202805,
-0.17853723466396332,
0.15191388130187988,
-0.1455254703760147,
-0.032587360590696335,
0.014678549952805042,
-0.08637823164463043,
0.005693607032299042,
0.10810600221157074,
-0.02643977291882038,
0.0941295176744461,
0.09555316716432571,
0.016295665875077248,
0.11225232481956482,
-0.04456339776515961,
-0.13011029362678528,
-0.012865182012319565,
-0.09335286915302277,
-0.07537991553544998,
-0.0031466407235711813,
0.1498716026544571,
-0.03807862848043442,
-0.11531389504671097,
-0.19169911742210388,
0.005153133533895016,
-0.17559204995632172,
-0.03215332329273224,
-0.06457369774580002,
-0.010502713732421398,
-0.01379061583429575,
-0.1060684472322464,
-0.02839687466621399,
0.0031733831856399775,
-0.10521868616342545,
-0.02674182318150997,
0.02451474778354168,
0.07379341125488281,
-0.09272908419370651,
0.03022477775812149,
0.07599029690027237,
0.06633209437131882,
0.17045408487319946,
0.12935452163219452,
-0.02048373967409134,
0.14116649329662323,
-0.15220317244529724,
-0.03280169144272804,
0.0009427535114809871,
-0.0004038145416416228,
0.033874887973070145,
0.1505042314529419,
-0.03889619931578636,
-0.008607588708400726,
-0.020384827628731728,
0.05438459292054176,
0.041318587958812714,
-0.12449812144041061,
-0.10502029210329056,
-0.020518874749541283,
-0.022749967873096466,
-0.02086039073765278,
-0.07232251018285751,
0.1361023187637329,
0.06574339419603348,
0.06077757477760315,
-0.056023791432380676,
-0.007710046600550413,
-0.07793040573596954,
0.0057066441513597965,
0.0388084352016449,
-0.0424177311360836,
0.005282138008624315,
-0.07634353637695312,
-0.06635686010122299,
-0.030199039727449417,
0.21628621220588684,
-0.06199687719345093,
-0.060989800840616226,
0.03162135183811188,
0.0038417764008045197,
0.03308137133717537,
-0.04059178754687309,
0.2867511212825775,
0.08177120983600616,
0.0038327910006046295,
-0.0546611063182354,
0.04287571832537651,
-0.030150311067700386,
-0.1651206612586975,
0.04409346356987953,
0.0011254828423261642,
0.01951662264764309,
0.020720265805721283,
0.1719353348016739,
-0.07672972232103348,
-0.0058965557254850864,
-0.029912194237113,
-0.010376310907304287,
-0.009541252627968788,
-0.0823773592710495,
0.06319744884967804,
0.17760539054870605,
-0.04174766317009926,
0.040246814489364624,
0.0097167007625103,
-0.02535620704293251,
-0.06763311475515366,
-0.09547095745801926,
-0.03491155058145523,
-0.22772853076457977,
0.08032232522964478,
-0.0849682092666626,
0.07716004550457001,
0.07186947762966156,
0.049947429448366165,
-0.0544382706284523,
-0.018575817346572876,
-0.0305961761623621,
-0.10625121742486954,
0.033605169504880905,
-0.04908708110451698,
-0.00657192338258028,
-0.07885058224201202,
-0.005409369710832834,
0.05622599273920059,
-0.009039369411766529,
-0.09725195914506912,
0.07902403920888901,
-0.017104359343647957,
0.0037042242474853992,
-0.09753414988517761,
-0.04278455674648285,
-0.04590320959687233,
0.05705152451992035,
0.01195472851395607,
0.20955513417720795,
0.028072141110897064,
0.006473785266280174,
0.05006621032953262,
0.2173149585723877,
0.016637854278087616,
-0.09863844513893127,
-0.03756120800971985,
-0.04482755437493324,
-0.0677444189786911,
0.09623656421899796,
-0.06685104221105576,
-0.008962667547166348,
-0.08123218268156052,
0.2735544741153717,
0.2374894618988037,
-0.11094827204942703,
0.019687658175826073,
-0.03578954190015793,
0.02015787735581398,
0.07132156193256378,
0.1183646097779274,
0.06589875370264053,
0.23932084441184998,
-0.0998615026473999,
-0.01087156031280756,
0.016405850648880005,
0.031369853764772415,
-0.10064921528100967,
0.09911375492811203,
0.007362354546785355,
-0.07306338846683502,
-0.05946015194058418,
0.08950813859701157,
-0.2732202708721161,
0.0930723175406456,
-0.14757384359836578,
-0.10953216999769211,
-0.020834535360336304,
-0.09009704738855362,
0.07763514667749405,
0.09065090864896774,
0.04967472329735756,
-0.0642826184630394,
0.006599227897822857,
0.04053410142660141,
-0.008172399364411831,
-0.23583820462226868,
-0.14338919520378113,
0.08489981293678284,
0.017710460349917412,
0.12676392495632172,
-0.01638556644320488,
0.04484276846051216,
0.054251447319984436,
0.0340290293097496,
-0.060389064252376556,
0.03883247449994087,
0.015149110928177834,
-0.028537703678011894,
-0.13822069764137268,
-0.0722849890589714,
0.056656915694475174,
0.0006679497309960425,
0.039997443556785583,
-0.04630441591143608,
0.03622330725193024,
0.1255313903093338,
-0.011518389917910099,
-0.06588200479745865,
0.07733501493930817,
-0.08137502521276474,
0.08796120434999466,
-0.014276611618697643,
-0.03763357549905777,
-0.05105185508728027,
-0.011248454451560974,
0.0375269278883934,
0.013152849860489368,
-0.03255810961127281,
-0.017475677654147148,
-0.04547153413295746,
-0.0403439998626709,
0.20962369441986084,
-0.021083643659949303,
-0.1515541970729828,
0.0013398361625149846,
-0.0898812934756279,
0.13014569878578186,
-0.020069818943738937,
0.14680729806423187,
0.1998165100812912,
0.02358233742415905,
0.009370217099785805,
-0.05339355394244194,
0.033524639904499054,
0.006669948808848858,
-0.04265107214450836,
-0.08166827261447906
] |
null | null | null | ## Exporting LoRas for use in other tools
Documentation on how to use Eden concepts in Automatic1111 or ComfyUI is here:
https://docs.eden.art/docs/guides/concepts/#exporting-loras-for-use-in-other-tools
| {} | null | eden-art/Elvara | [
"region:us"
] | 2024-02-06T07:40:44+00:00 | [] | [] | TAGS
#region-us
| ## Exporting LoRas for use in other tools
Documentation on how to use Eden concepts in Automatic1111 or ComfyUI is here:
URL
| [
"## Exporting LoRas for use in other tools\nDocumentation on how to use Eden concepts in Automatic1111 or ComfyUI is here:\nURL"
] | [
"TAGS\n#region-us \n",
"## Exporting LoRas for use in other tools\nDocumentation on how to use Eden concepts in Automatic1111 or ComfyUI is here:\nURL"
] | [
6,
31
] | [
"passage: TAGS\n#region-us \n## Exporting LoRas for use in other tools\nDocumentation on how to use Eden concepts in Automatic1111 or ComfyUI is here:\nURL"
] | [
-0.04433843493461609,
0.045720089226961136,
-0.00047354400157928467,
0.030475227162241936,
0.06009476259350777,
0.07998326420783997,
0.15519832074642181,
0.06912561506032944,
0.10809514671564102,
0.012642289511859417,
0.1122450977563858,
0.22898027300834656,
-0.00028577446937561035,
0.05475001782178879,
0.044760413467884064,
-0.13118614256381989,
0.038832277059555054,
0.0040464517660439014,
-0.05706696957349777,
0.006891309283673763,
0.05449127405881882,
-0.01229453831911087,
0.12554284930229187,
0.002862817607820034,
-0.142470121383667,
-0.030325444415211678,
-0.07109984010457993,
-0.020695682615041733,
0.15232129395008087,
0.01525056641548872,
0.11346510052680969,
-0.05962107330560684,
0.06726023554801941,
-0.2382640540599823,
0.006167016923427582,
-0.04446452483534813,
-0.08604331314563751,
0.038463905453681946,
0.032531023025512695,
0.008865173906087875,
0.08055979013442993,
0.022647539153695107,
-0.040747568011283875,
0.029510561376810074,
-0.1292288601398468,
-0.09273699671030045,
-0.01919589936733246,
-0.20647749304771423,
0.041826311498880386,
0.07085409760475159,
0.06684459000825882,
0.10237656533718109,
0.0012950926320627332,
0.08269114792346954,
0.07960638403892517,
-0.2375565767288208,
0.01256447285413742,
0.056798405945301056,
-0.010187654756009579,
0.1284167617559433,
0.019812004640698433,
0.095412977039814,
0.10692135989665985,
0.007207931485027075,
-0.07240255922079086,
-0.09751627594232559,
-0.2322804033756256,
-0.060488536953926086,
-0.05347404256463051,
-0.046161483973264694,
0.32319942116737366,
0.01947224885225296,
-0.04151201993227005,
-0.0005828229477629066,
-0.08579631894826889,
-0.02140403538942337,
-0.06669867038726807,
0.037897855043411255,
0.04411221668124199,
0.09293074905872345,
0.1419050693511963,
-0.13018806278705597,
-0.15846356749534607,
-0.04179900139570236,
-0.014519616030156612,
0.10167136788368225,
0.024988526478409767,
0.15101222693920135,
-0.0903594121336937,
-0.008592082187533379,
-0.14549174904823303,
-0.06266815960407257,
0.02128669060766697,
-0.009600776247680187,
0.1391984522342682,
0.06077577918767929,
0.031002547591924667,
-0.028631221503019333,
0.19688726961612701,
0.07332884520292282,
0.29705730080604553,
-0.028248770162463188,
-0.0008367539849132299,
0.10764946788549423,
0.10136397182941437,
0.031696245074272156,
0.1471766084432602,
-0.0895332470536232,
0.09939420223236084,
-0.08287346363067627,
0.06309664994478226,
-0.057791873812675476,
-0.15300920605659485,
-0.050576165318489075,
-0.022225933149456978,
0.1351950764656067,
0.0029614015948027372,
-0.10680586844682693,
-0.08497141301631927,
0.005826888140290976,
0.12139888852834702,
-0.07320283353328705,
-0.04358486831188202,
-0.10843203961849213,
0.04476933181285858,
0.06713709235191345,
0.029085438698530197,
0.061308059841394424,
0.038029544055461884,
-0.14176931977272034,
-0.06382955610752106,
-0.0025079569313675165,
-0.02186186984181404,
0.010220520198345184,
-0.04657070338726044,
-0.09952163696289062,
0.1288270503282547,
-0.170726478099823,
-0.0846080482006073,
-0.00939324963837862,
0.09926830977201462,
0.04491020739078522,
0.051859285682439804,
-0.05294939503073692,
0.02979426272213459,
-0.008650673553347588,
-0.01325561385601759,
-0.08244112879037857,
-0.0654679611325264,
-0.00042870640754699707,
-0.030844638124108315,
0.032902792096138,
-0.2546124756336212,
0.020548837259411812,
-0.12120025604963303,
-0.03200944885611534,
0.0027377838268876076,
-0.011877001263201237,
-0.05287269875407219,
0.161833718419075,
-0.12214702367782593,
0.05871191248297691,
-0.003084423951804638,
0.05154859274625778,
-0.008803712204098701,
0.11371197551488876,
-0.19819366931915283,
0.047961361706256866,
0.1567688137292862,
-0.1455976963043213,
-0.1926165521144867,
0.15881820023059845,
0.0036157630383968353,
0.11565403640270233,
0.057833340018987656,
0.21578609943389893,
0.09015526622533798,
-0.08982794731855392,
0.10948847234249115,
0.10069915652275085,
0.007531811483204365,
-0.1669207215309143,
0.011186666786670685,
0.02176087349653244,
-0.012201394885778427,
0.05147595703601837,
-0.021764572709798813,
0.056990694254636765,
0.004992435220628977,
-0.116025410592556,
-0.03727976232767105,
-0.1064009889960289,
-0.13632968068122864,
0.005449695512652397,
-0.002213813830167055,
-0.03347337245941162,
0.08037640899419785,
-0.255450040102005,
0.10399901866912842,
-0.03684118390083313,
0.013522331602871418,
0.012947450391948223,
0.067631796002388,
-0.05744153633713722,
-0.030038923025131226,
-0.043572381138801575,
-0.12639829516410828,
0.058672260493040085,
0.07737462967634201,
0.0939059928059578,
0.17336401343345642,
0.02350931614637375,
0.06391274929046631,
0.08630427718162537,
-0.02257007360458374,
-0.04891660809516907,
-0.046098098158836365,
-0.05303635075688362,
-0.013125800527632236,
0.05845243111252785,
-0.07933436334133148,
-0.0477922223508358,
-0.09011168032884598,
-0.04113229364156723,
-0.007237908896058798,
-0.0400346964597702,
0.05666050687432289,
-0.0020247262436896563,
-0.020687120035290718,
-0.009574482217431068,
-0.11348255723714828,
-0.006944268010556698,
0.07196864485740662,
-0.04367543011903763,
-0.18159128725528717,
0.06113710626959801,
-0.09439893811941147,
0.049932725727558136,
0.12062124907970428,
-0.1557563990354538,
-0.0031203548423945904,
-0.046046141535043716,
0.03498651459813118,
0.04326348006725311,
-0.04754815995693207,
-0.024400530382990837,
0.21610808372497559,
-0.02639778144657612,
0.062050074338912964,
-0.02500511147081852,
0.09635340422391891,
-0.028268659487366676,
-0.09092239290475845,
-0.07349438965320587,
0.048879507929086685,
-0.02844330109655857,
-0.11916031688451767,
0.05331830307841301,
0.2548058331012726,
0.009456834755837917,
0.09176382422447205,
-0.039839643985033035,
-0.029517557471990585,
-0.014833444729447365,
0.11720269173383713,
0.045724380761384964,
0.057815901935100555,
-0.07315077632665634,
-0.06230960413813591,
0.01912537030875683,
0.003362814197316766,
0.017168521881103516,
-0.16912256181240082,
-0.06059691309928894,
-0.010752328671514988,
-0.021023614332079887,
0.16543658077716827,
0.005997363943606615,
-0.11231280863285065,
0.017261601984500885,
-0.04193988814949989,
-0.11772774159908295,
0.02725001610815525,
-0.03795741870999336,
0.008380708284676075,
0.1406129002571106,
-0.13786377012729645,
-0.06919067353010178,
-0.1889580637216568,
0.04974452406167984,
-0.008745449595153332,
-0.0015079820295795798,
0.10244690626859665,
-0.09919951856136322,
-0.13137191534042358,
-0.02174617536365986,
-0.03333283215761185,
-0.04610222578048706,
-0.07556098699569702,
0.022460946813225746,
0.05486924573779106,
-0.13076551258563995,
-0.1704493761062622,
-0.04443753510713577,
-0.018432382494211197,
-0.09417583793401718,
0.09685434401035309,
-0.0997503325343132,
0.1411323845386505,
-0.010019953362643719,
0.0652056336402893,
-0.017075158655643463,
0.08022399246692657,
0.26073065400123596,
-0.06249592825770378,
-0.0020055335480719805,
0.27167537808418274,
-0.04389049485325813,
0.08615709096193314,
0.09429879486560822,
0.08415935933589935,
-0.211930513381958,
0.0022861803881824017,
-0.10823482275009155,
-0.13492080569267273,
-0.21380464732646942,
-0.10615858435630798,
-0.12173449993133545,
-0.05280851572751999,
0.059740543365478516,
0.05357161909341812,
0.09410455077886581,
0.1829243153333664,
0.059625882655382156,
0.15032057464122772,
-0.017786435782909393,
0.06434947997331619,
-0.07219313085079193,
0.027950601652264595,
0.0031125973910093307,
-0.10677524656057358,
-0.09931692481040955,
0.09079386293888092,
0.25386837124824524,
0.22916698455810547,
0.18811145424842834,
0.3258594274520874,
0.06527011841535568,
0.11509647965431213,
0.04932282119989395,
0.15715926885604858,
-0.04249927029013634,
0.0006276392959989607,
-0.08597801625728607,
-0.04454759135842323,
-0.11615798622369766,
0.13142231106758118,
0.028850965201854706,
-0.11904729157686234,
0.0968024805188179,
-0.08242344856262207,
0.013460773043334484,
0.07133956998586655,
0.1249711662530899,
-0.16033852100372314,
0.13460089266300201,
0.11830779165029526,
0.045924749225378036,
-0.0722290500998497,
0.11284724622964859,
0.05378531664609909,
-0.028219042345881462,
-0.021643180400133133,
0.023189306259155273,
0.11865269392728806,
-0.17477037012577057,
-0.029196463525295258,
-0.21092639863491058,
-0.027888966724276543,
0.05944688245654106,
0.05748385190963745,
-0.0882025808095932,
0.2565796673297882,
0.04548555985093117,
-0.04749630391597748,
-0.08265742659568787,
0.003985634073615074,
0.0807216539978981,
0.03456588834524155,
0.1490287184715271,
0.009230068884789944,
-0.27603787183761597,
-0.12767232954502106,
-0.09948618710041046,
0.06802312284708023,
0.0869322344660759,
-0.05563376471400261,
-0.08915433287620544,
-0.0923846960067749,
0.029343077912926674,
-0.05065131187438965,
0.04431116208434105,
-0.11318527907133102,
-0.1586136370897293,
-0.017073802649974823,
0.22452707588672638,
0.02854950912296772,
0.02921243943274021,
0.08075454831123352,
0.04102405533194542,
0.11768347769975662,
0.04787959158420563,
-0.03838489577174187,
-0.07473257929086685,
-0.16151340305805206,
0.0800650492310524,
-0.07848789542913437,
0.0830262154340744,
-0.07976096123456955,
-0.018323425203561783,
-0.07901298999786377,
-0.1466776728630066,
0.04952969029545784,
-0.07913156598806381,
0.07594480365514755,
-0.0398239903151989,
0.03662985563278198,
0.03377999737858772,
-0.05206592008471489,
0.09767986088991165,
0.043211955577135086,
-0.11555317789316177,
-0.09740472584962845,
-0.0019275665981695056,
0.13224183022975922,
-0.02951066941022873,
0.08853008598089218,
-0.1889900118112564,
-0.0809108167886734,
-0.09280940145254135,
-0.07828979194164276,
0.3056708872318268,
0.2289346605539322,
-0.04763748124241829,
-0.04046602547168732,
0.22480736672878265,
-0.05928864702582359,
-0.2550167441368103,
-0.15209494531154633,
-0.03601517155766487,
-0.003408471355214715,
0.10284291207790375,
-0.11220849305391312,
0.2014131247997284,
0.13420197367668152,
-0.03570098429918289,
0.2771279513835907,
-0.07361812144517899,
-0.022754747420549393,
0.12137997895479202,
0.2021537572145462,
0.24221237003803253,
-0.13860997557640076,
0.01907336339354515,
-0.07071713358163834,
-0.09990965574979782,
0.10107512772083282,
0.08088795840740204,
0.08055137097835541,
0.03707817196846008,
0.14914996922016144,
0.03508137911558151,
0.007259644567966461,
0.2230025678873062,
0.019177282229065895,
0.09754124283790588,
-0.019563740119338036,
0.002023818902671337,
0.028801865875720978,
-0.029165232554078102,
0.0704037994146347,
-0.16438017785549164,
-0.015887277200818062,
-0.021869463846087456,
0.05852537602186203,
0.019583525136113167,
0.08915113657712936,
0.021787814795970917,
-0.10000763833522797,
-0.03896050527691841,
-0.05721646919846535,
-0.02572944015264511,
0.09767455607652664,
0.3132671117782593,
0.018424024805426598,
0.03195596858859062,
0.08630215376615524,
0.030713502317667007,
0.06630376726388931,
0.026465240865945816,
-0.07766943424940109,
-0.04500589519739151,
0.0531456284224987,
-0.1619589924812317,
0.024344660341739655,
0.08087703585624695,
-0.025133362039923668,
0.018239639699459076,
-0.00729906652122736,
-0.05704394727945328,
0.07882632315158844,
0.06920463591814041,
0.004963916260749102,
-0.05729948729276657,
0.008607710711658001,
0.10670829564332962,
0.08609560132026672,
0.02682562544941902,
0.1747404783964157,
-0.004955344833433628,
-0.033218760043382645,
-0.053558237850666046,
0.02226593904197216,
-0.1261436939239502,
-0.027129964902997017,
-0.0037195165641605854,
0.008087234571576118,
-0.08157002180814743,
0.024356381967663765,
0.005504644475877285,
-0.02123669721186161,
-0.031372133642435074,
0.029017552733421326,
-0.021535124629735947,
-0.16209867596626282,
0.08736726641654968,
0.09407568722963333,
-0.14305831491947174,
-0.004958394914865494,
0.03073323704302311,
0.05721001327037811,
-0.15323406457901,
0.0300708319991827,
0.0757087916135788,
0.0027494006790220737,
0.00194677640683949,
-0.039665814489126205,
0.046624429523944855,
-0.07484503090381622,
-0.1579275131225586,
0.056249335408210754,
-0.1182822734117508,
-0.18041273951530457,
0.006753907073289156,
0.06312363594770432,
-0.04456675425171852,
-0.056523844599723816,
-0.09002676606178284,
0.019211329519748688,
0.07765380293130875,
0.09820868074893951,
-0.0652141198515892,
0.04620560258626938,
0.018545091152191162,
-0.06609894335269928,
-0.027702782303094864,
0.01930282823741436,
-0.13850508630275726,
0.015256228856742382,
-0.0053676459938287735,
0.05135096609592438,
0.01732400432229042,
-0.12570540606975555,
0.07260168343782425,
-0.01743260584771633,
-0.011471051722764969,
0.01582249440252781,
0.007894204929471016,
0.1725761741399765,
-0.18268154561519623,
0.03670097887516022,
0.14401370286941528,
0.0887673944234848,
0.06930580735206604,
0.13572649657726288,
-0.10086996853351593,
-0.03529210761189461,
0.04182075336575508,
-0.039647795259952545,
0.16314885020256042,
-0.10542714595794678,
-0.07277381420135498,
-0.045855842530727386,
-0.2422468066215515,
-0.06571492552757263,
-0.06534810364246368,
0.16194038093090057,
0.025087246671319008,
0.10548178851604462,
0.048734188079833984,
0.13628610968589783,
0.0421445295214653,
0.04198283702135086,
-0.0009009785135276616,
-0.1272924244403839,
-0.09323607385158539,
-0.14091001451015472,
-0.04535719379782677,
-0.051776230335235596,
0.18409767746925354,
-0.07441477477550507,
0.057391341775655746,
-0.050427086651325226,
0.04230416193604469,
-0.05118660628795624,
-0.055851005017757416,
0.1726982593536377,
0.08784762024879456,
0.06313201040029526,
-0.21310657262802124,
0.047927048057317734,
0.0389467217028141,
-0.08518927544355392,
-0.019237849861383438,
0.016446860507130623,
0.0038429382257163525,
0.08183915168046951,
0.045075755566358566,
0.03666992858052254,
-0.09467121213674545,
-0.2716463506221771,
0.12775015830993652,
0.05218576639890671,
0.051206815987825394,
0.0930371955037117,
0.13817143440246582,
-0.019601289182901382,
0.036981429904699326,
0.04219929873943329,
-0.024763189256191254,
-0.1384345293045044,
-0.16157369315624237,
-0.05968788266181946,
-0.03657170757651329,
0.01252133771777153,
-0.043104980140924454,
0.0041328188963234425,
0.08965103328227997,
0.012408087961375713,
0.03177161514759064,
0.0881163626909256,
-0.035729702562093735,
-0.0353059358894825,
-0.005539592821151018,
0.027690928429365158,
-0.00352687225677073,
-0.022278590127825737,
-0.06858806312084198,
-0.13300766050815582,
-0.004843530710786581,
-0.04542946070432663,
-0.026485931128263474,
-0.03638312220573425,
0.0512576624751091,
-0.1002402976155281,
-0.11889449506998062,
-0.11009000241756439,
-0.011190476827323437,
-0.0706157460808754,
-0.05798827111721039,
0.0692591443657875,
-0.07900942862033844,
0.06195252761244774,
0.08887068927288055,
-0.0199394840747118,
0.06673236936330795,
0.003929920494556427,
0.0944267213344574,
-0.12760625779628754,
0.04468938708305359,
-0.061677515506744385,
0.02684980444610119,
-0.009246641770005226,
0.15672938525676727,
0.19249823689460754,
-0.08679144829511642,
0.017688389867544174,
-0.010758941061794758,
0.010621597059071064,
-0.050446584820747375,
0.14828504621982574,
0.014540954492986202,
0.19881783425807953,
-0.08510805666446686,
-0.020967422053217888,
-0.07866624742746353,
0.03629322350025177,
-0.20212338864803314,
-0.09103287011384964,
-0.06276603788137436,
-0.004034777171909809,
-0.025419078767299652,
0.14051681756973267,
-0.056109748780727386,
-0.007874635979533195,
0.12759019434452057,
-0.11785557121038437,
-0.0953422263264656,
-0.06017377972602844,
0.07429029047489166,
0.003203506814315915,
0.07549522072076797,
-0.0985058844089508,
-0.02728458121418953,
0.0003510818933136761,
0.0013664837460964918,
-0.24580009281635284,
-0.06417541950941086,
-0.007523628883063793,
-0.10273639112710953,
0.12329849600791931,
0.032368991523981094,
-0.033683061599731445,
0.03036896139383316,
-0.013998684473335743,
0.012439553625881672,
0.11505819857120514,
0.02644996903836727,
0.006391195114701986,
-0.00769032770767808,
-0.024079909548163414,
-0.10549765080213547,
-0.020136617124080658,
0.06407961994409561,
-0.168318510055542,
0.0014906266005709767,
0.23268362879753113,
0.008850977756083012,
0.004184948280453682,
0.044439177960157394,
-0.15888074040412903,
0.12268789857625961,
0.014158028177917004,
-0.00007685019227210432,
-0.09443701058626175,
-0.004671953618526459,
0.04941612482070923,
0.15472620725631714,
-0.05525188893079758,
-0.13595391809940338,
-0.006282376125454903,
-0.12508809566497803,
0.08519342541694641,
-0.021866852417588234,
-0.046936821192502975,
-0.06261269748210907,
-0.14761024713516235,
-0.012050008401274681,
-0.16500867903232574,
-0.0007083576056174934,
0.20348358154296875,
-0.0973268523812294,
0.03923256695270538,
-0.16589230298995972,
0.056608643382787704,
-0.04090796038508415,
-0.04958365485072136,
-0.13029515743255615
] |
null | null | transformers |
# Definition
[phi-2] for [P]ersonal [I]dentifiable [I]nformation with [B]anking [B]anking [I]nsurance Dataset
# How to use model
## Load model and tokenizer
```
import torch
from transformers import AutoModelForCausalLM, BitsAndBytesConfig, AutoTokenizer
torch.set_default_device("cuda")
model_name = "dcipheranalytics/phi-2-pii-bbi"
quantization_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_compute_dtype=torch.bfloat16,
bnb_4bit_quant_type="nf4",
)
model = AutoModelForCausalLM.from_pretrained(
model_name,
device_map="auto",
# torch_dtype="auto",
torch_dtype=torch.bfloat16,
trust_remote_code=True,
quantization_config=quantization_config,
)
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
```
## Call generate method
```
def generate(msg: str, max_new_tokens = 300, temperature=0.3):
chat_template = "<|im_start|>user\n{msg}<|im_end|><|im_start|>assistant\n"
prompt = chat_template.format(msg=msg)
with torch.no_grad():
token_ids = tokenizer.encode(prompt, add_special_tokens=False, return_tensors="pt")
output_ids = model.generate(
token_ids.to(model.device),
max_new_tokens=max_new_tokens,
do_sample=True,
temperature=temperature,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=tokenizer.eos_token_id,
)
output = tokenizer.decode(output_ids[0][token_ids.size(1):-1]).strip()
return output
instruction_template = "List the personally identifiable information in the given text below.\nText:########\n{text}\n########"
text_with_pii = "My passport number is 123456789."
generate(instruction_template.format(text=text_with_pii))
```
## Batch predictions
```
from transformers import TextGenerationPipeline
def get_prompt(text):
instruction_template = "List the personally identifiable information in the given text below.\nText:########\n{text}\n########"
msg = instruction_template.format(text=text)
chat_template = "<|im_start|>user\n{msg}<|im_end|><|im_start|>assistant\n"
prompt = chat_template.format(msg=msg)
return prompt
generator = TextGenerationPipeline(
model=model,
tokenizer=tokenizer,
max_new_tokens=300,
do_sample=True,
temperature=0.3,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=tokenizer.eos_token_id,
)
texts = ["My passport number is 123456789.",
"My name is John Smith.",
]
prompts = list(map(get_prompt, texts))
outputs = generator(prompts,
return_full_text=False,
batch_size=2)
```
# Train Data
GPT4 generated customer service conversations.
1. 100 unique banking topics, 8 examples per each,
2. New 100 banking topics, 4 examples per each,
3. 100 insurance topics, 4 examples per each.
# Evaluation Results
## Average
```
precision 0.836223
recall 0.781132
f1 0.801837
```
## Per topic:

## On TAB test split:
```
precision 0.506118
recall 0.350976
f1 0.391614
``` | {"language": ["en"]} | text-generation | dcipheranalytics/phi-2-pii-bbi | [
"transformers",
"safetensors",
"phi",
"text-generation",
"custom_code",
"en",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T07:42:38+00:00 | [] | [
"en"
] | TAGS
#transformers #safetensors #phi #text-generation #custom_code #en #autotrain_compatible #endpoints_compatible #region-us
|
# Definition
[phi-2] for [P]ersonal [I]dentifiable [I]nformation with [B]anking [B]anking [I]nsurance Dataset
# How to use model
## Load model and tokenizer
## Call generate method
## Batch predictions
# Train Data
GPT4 generated customer service conversations.
1. 100 unique banking topics, 8 examples per each,
2. New 100 banking topics, 4 examples per each,
3. 100 insurance topics, 4 examples per each.
# Evaluation Results
## Average
## Per topic:
!image/png
## On TAB test split:
| [
"# Definition\n\n[phi-2] for [P]ersonal [I]dentifiable [I]nformation with [B]anking [B]anking [I]nsurance Dataset",
"## Call generate method",
"## Batch predictions",
"# Train Data\n\nGPT4 generated customer service conversations. \n1. 100 unique banking topics, 8 examples per each,\n2. New 100 banking topics, 4 examples per each,\n3. 100 insurance topics, 4 examples per each.",
"# Evaluation Results",
"## Average",
"## Per topic:\n!image/png",
"## On TAB test split:"
] | [
"TAGS\n#transformers #safetensors #phi #text-generation #custom_code #en #autotrain_compatible #endpoints_compatible #region-us \n",
"# Definition\n\n[phi-2] for [P]ersonal [I]dentifiable [I]nformation with [B]anking [B]anking [I]nsurance Dataset",
"## Call generate method",
"## Batch predictions",
"# Train Data\n\nGPT4 generated customer service conversations. \n1. 100 unique banking topics, 8 examples per each,\n2. New 100 banking topics, 4 examples per each,\n3. 100 insurance topics, 4 examples per each.",
"# Evaluation Results",
"## Average",
"## Per topic:\n!image/png",
"## On TAB test split:"
] | [
44,
43,
4,
5,
53,
4,
3,
8,
6
] | [
"passage: TAGS\n#transformers #safetensors #phi #text-generation #custom_code #en #autotrain_compatible #endpoints_compatible #region-us \n# Definition\n\n[phi-2] for [P]ersonal [I]dentifiable [I]nformation with [B]anking [B]anking [I]nsurance Dataset## Call generate method## Batch predictions# Train Data\n\nGPT4 generated customer service conversations. \n1. 100 unique banking topics, 8 examples per each,\n2. New 100 banking topics, 4 examples per each,\n3. 100 insurance topics, 4 examples per each.# Evaluation Results## Average## Per topic:\n!image/png## On TAB test split:"
] | [
-0.12731735408306122,
-0.03894650191068649,
-0.0030404776334762573,
0.020259719341993332,
0.14718995988368988,
-0.02025759220123291,
0.1381717324256897,
0.07237771898508072,
0.0015043173916637897,
0.08610434830188751,
0.1060556173324585,
0.1206495612859726,
0.03225032985210419,
0.30965906381607056,
-0.06942097842693329,
-0.16741128265857697,
0.08213035017251968,
0.0026240795850753784,
0.1425410360097885,
0.13246549665927887,
0.12342750281095505,
-0.0989311933517456,
0.07649843394756317,
-0.03149803727865219,
-0.1433057337999344,
-0.023530008271336555,
0.045491017401218414,
-0.12261955440044403,
0.13201795518398285,
0.024609224870800972,
0.11076324433088303,
-0.007603875827044249,
0.020476100966334343,
-0.16706565022468567,
0.016470326110720634,
-0.004821428097784519,
-0.02532896399497986,
0.02578120492398739,
0.08397313952445984,
-0.04313559830188751,
0.02620364911854267,
0.05124306306242943,
0.10605180263519287,
0.028943970799446106,
-0.14373153448104858,
-0.09114691615104675,
-0.09553161263465881,
0.05137884244322777,
0.14947235584259033,
0.08587473630905151,
-0.03671934828162193,
0.08205955475568771,
-0.1731579601764679,
0.07653795927762985,
0.17081648111343384,
-0.33149874210357666,
0.005096756387501955,
0.054142825305461884,
0.02277938276529312,
0.02872287482023239,
-0.08012694120407104,
-0.04260789230465889,
0.04329529404640198,
0.03828372061252594,
0.06797073036432266,
-0.03740108758211136,
-0.06171523407101631,
0.04436304792761803,
-0.19970040023326874,
-0.027620507404208183,
0.15517336130142212,
0.05489061027765274,
-0.08063371479511261,
-0.16518093645572662,
-0.04947550967335701,
-0.07434944808483124,
-0.0628390982747078,
-0.02270263060927391,
-0.04838947579264641,
-0.07045553624629974,
-0.06105222553014755,
0.12496131658554077,
-0.05987807363271713,
-0.0687783733010292,
-0.13889692723751068,
0.04910944402217865,
0.02457248978316784,
0.03627096489071846,
-0.0010719172423705459,
0.06970299780368805,
-0.11531959474086761,
-0.03510400652885437,
-0.04879331961274147,
-0.05958893150091171,
-0.06651533395051956,
-0.0519137978553772,
-0.00555784534662962,
-0.023119140416383743,
0.08860412240028381,
0.14671829342842102,
-0.04301568865776062,
0.01658880151808262,
-0.043368980288505554,
0.036334022879600525,
0.02151135727763176,
0.09424412250518799,
0.007091079838573933,
-0.09236138314008713,
-0.03958258405327797,
0.0647740587592125,
-0.018795380368828773,
0.014981402084231377,
-0.06474194675683975,
0.07989662140607834,
0.0914824903011322,
0.06941346079111099,
-0.06201312690973282,
0.05356549844145775,
-0.08256174623966217,
-0.0001545087288832292,
0.13432858884334564,
-0.12814441323280334,
-0.006550335790961981,
-0.008832057937979698,
0.012252422980964184,
-0.10298161208629608,
-0.03463202342391014,
-0.000008335659913427662,
-0.039920054376125336,
0.08949577808380127,
-0.05160069465637207,
-0.01723865419626236,
-0.06533211469650269,
-0.07041102647781372,
0.05383804440498352,
-0.062382448464632034,
0.00032539982930757105,
-0.1287631094455719,
-0.23951008915901184,
-0.015432360582053661,
-0.041476815938949585,
-0.049151819199323654,
-0.04922371730208397,
-0.01898191310465336,
0.008993935771286488,
-0.0635627806186676,
-0.0635935515165329,
0.05434205010533333,
-0.08513136208057404,
0.048588827252388,
0.02334011346101761,
0.05982104316353798,
0.015529511496424675,
0.015574133954942226,
-0.10972807556390762,
0.009434765204787254,
-0.13997410237789154,
0.03298238664865494,
-0.11185792833566666,
0.054066646844148636,
-0.0608348473906517,
-0.015621312893927097,
-0.028290653601288795,
0.018748566508293152,
0.0038464341778308153,
0.13410504162311554,
-0.046077046543359756,
-0.08570569008588791,
0.263671875,
-0.1877221018075943,
-0.10681328922510147,
0.11865906417369843,
-0.04474575072526932,
-0.003810698864981532,
0.10250218212604523,
0.19580069184303284,
0.020243944600224495,
-0.06915369629859924,
-0.010051420889794827,
0.04251500219106674,
-0.010939028114080429,
-0.07555191218852997,
0.09330294281244278,
-0.057472649961709976,
-0.07964704930782318,
0.05461408942937851,
0.051484961062669754,
-0.00001178182355943136,
-0.12186023592948914,
-0.0005787934060208499,
0.012433076277375221,
-0.0716441199183464,
0.0009217049228027463,
0.02459564432501793,
0.027450082823634148,
-0.067426897585392,
-0.08484002947807312,
0.001085606636479497,
0.04052430018782616,
-0.002668839879333973,
-0.017856605350971222,
-0.12551958858966827,
0.2497030347585678,
-0.029080962762236595,
0.008017236366868019,
-0.13034683465957642,
-0.03743710368871689,
0.026905033737421036,
0.027885518968105316,
-0.00863021332770586,
0.05642174929380417,
0.05925063416361809,
0.037213750183582306,
-0.059511926025152206,
-0.0021311002783477306,
0.08911357820034027,
0.00874597579240799,
-0.09478059411048889,
-0.12170001119375229,
0.0002856722567230463,
-0.11606290936470032,
0.09716282039880753,
-0.14893239736557007,
0.0040398212149739265,
0.03237483650445938,
0.06977181136608124,
0.06278833001852036,
-0.03759651258587837,
-0.01922561787068844,
0.04046754911541939,
-0.02761935442686081,
-0.006553381681442261,
0.020260736346244812,
0.004629897885024548,
-0.02497224509716034,
0.09336483478546143,
-0.19829486310482025,
0.2593555152416229,
0.10586017370223999,
-0.10756876319646835,
-0.07024877518415451,
-0.10413291305303574,
-0.07443288713693619,
0.032603919506073,
-0.14026756584644318,
-0.058173999190330505,
-0.013226741924881935,
0.0006652289885096252,
0.1263934075832367,
-0.0660209208726883,
-0.02542725019156933,
0.019698474556207657,
-0.02777373418211937,
-0.006728343665599823,
0.12984998524188995,
-0.004854994360357523,
-0.057232458144426346,
0.09349215775728226,
0.11229569464921951,
0.05659852549433708,
0.05207172408699989,
-0.037136148661375046,
-0.07794229686260223,
0.021839190274477005,
0.05981021746993065,
0.009094314649701118,
0.1136268600821495,
-0.14365775883197784,
-0.029954129830002785,
0.09579312056303024,
0.017159556970000267,
0.03254396468400955,
-0.09701661765575409,
-0.029912488535046577,
-0.00733193289488554,
-0.03341374173760414,
0.006762286648154259,
0.03937868773937225,
0.016546916216611862,
0.11428684741258621,
0.03108377382159233,
-0.11946958303451538,
0.05517932027578354,
0.001628048368729651,
-0.10894820094108582,
0.23687614500522614,
-0.023128045722842216,
-0.212659552693367,
-0.12345505505800247,
-0.16309034824371338,
0.05140514299273491,
0.023732393980026245,
0.0829726979136467,
-0.06844192743301392,
-0.07148231565952301,
-0.11450611799955368,
0.05623609572649002,
0.019550470635294914,
-0.02221747674047947,
0.029742339625954628,
-0.04403383284807205,
-0.03840743750333786,
-0.06524240970611572,
-0.05680535361170769,
-0.06047533452510834,
-0.06393538415431976,
0.11384069919586182,
-0.05782482773065567,
0.08964476734399796,
0.17374174296855927,
-0.007126688491553068,
0.03814021870493889,
-0.0957183912396431,
0.19939906895160675,
-0.0311959870159626,
-0.074336938560009,
0.14372187852859497,
-0.04866018146276474,
0.06101692095398903,
0.1669108122587204,
0.02748117968440056,
-0.08091386407613754,
-0.005677357781678438,
0.045744720846414566,
-0.0580846332013607,
-0.2283099889755249,
-0.09243680536746979,
-0.04250122234225273,
0.1067838966846466,
-0.0569288432598114,
0.05215844511985779,
0.11780687421560287,
0.061520010232925415,
0.038818664848804474,
-0.06751396507024765,
0.02862386405467987,
0.09987788647413254,
0.16205722093582153,
-0.016531571745872498,
0.13971947133541107,
-0.0232382919639349,
-0.05672275647521019,
0.09645155072212219,
-0.07773082703351974,
0.2946673631668091,
0.03859030082821846,
0.03174717724323273,
0.08054878562688828,
0.016815729439258575,
0.08371960371732712,
0.06857340037822723,
0.0032747506629675627,
-0.01950220763683319,
-0.017421161755919456,
-0.022224610671401024,
-0.04878583922982216,
0.020756611600518227,
-0.08046658337116241,
-0.047678373754024506,
-0.09927419573068619,
-0.0011673734989017248,
0.042500510811805725,
0.12829871475696564,
0.1485702097415924,
-0.29192954301834106,
-0.01863406039774418,
0.01110120303928852,
-0.15507890284061432,
-0.09474107623100281,
0.05600912868976593,
-0.0957828089594841,
-0.16534675657749176,
0.03831234201788902,
-0.046728890389204025,
0.08823450654745102,
-0.025170888751745224,
0.010300408117473125,
-0.048533130437135696,
-0.09867140650749207,
-0.05337778478860855,
0.10850849002599716,
-0.23918049037456512,
0.262725293636322,
-0.016777489334344864,
0.037474375218153,
-0.06616398692131042,
0.02203270234167576,
-0.08934356272220612,
0.05994775891304016,
0.1509241759777069,
-0.024111635982990265,
-0.007989361882209778,
-0.15881125628948212,
0.007624552119523287,
-0.006506725214421749,
0.049182794988155365,
0.0012050746008753777,
0.05853092297911644,
0.01951381377875805,
-0.005329241510480642,
-0.02196105197072029,
-0.10037259012460709,
-0.04212671518325806,
-0.10414319485425949,
0.03420369327068329,
-0.15799705684185028,
0.08801369369029999,
-0.08621546626091003,
-0.015519486740231514,
-0.17489250004291534,
0.12469932436943054,
0.009729980491101742,
-0.0652153268456459,
-0.09533877670764923,
0.03872322663664818,
0.015216710045933723,
-0.05256935581564903,
0.07650718837976456,
-0.0304363165050745,
0.047622889280319214,
0.030474001541733742,
-0.0660577043890953,
0.056161027401685715,
-0.10998495668172836,
-0.12587980926036835,
-0.07365618646144867,
0.1588277369737625,
-0.009549140930175781,
0.003910412546247244,
0.03334585577249527,
0.013829790987074375,
-0.044199492782354355,
-0.0740310475230217,
0.005427646916359663,
-0.03780977800488472,
0.13528968393802643,
0.10225173830986023,
-0.04133160039782524,
-0.11464790254831314,
-0.053033992648124695,
-0.06721866130828857,
0.1613140106201172,
0.24854227900505066,
-0.048128314316272736,
0.07111524790525436,
0.19627267122268677,
-0.011217559687793255,
-0.20259644091129303,
-0.06506676971912384,
-0.07996077835559845,
0.020750053226947784,
-0.0026738096494227648,
-0.032705143094062805,
0.17770452797412872,
0.09530236572027206,
-0.06197004392743111,
0.0800672248005867,
-0.16536502540111542,
-0.1040709912776947,
0.20638908445835114,
0.03306455537676811,
0.21768105030059814,
-0.15766438841819763,
-0.06432515382766724,
-0.010833276435732841,
-0.12666428089141846,
0.0617472268640995,
-0.1273876428604126,
0.037082962691783905,
-0.040119148790836334,
0.058255814015865326,
0.0038180937990546227,
-0.06975915282964706,
0.09020331501960754,
0.02471993863582611,
0.02433846704661846,
-0.08329463750123978,
-0.007650910876691341,
0.06423749029636383,
-0.03462005406618118,
0.04897230491042137,
-0.014635113999247551,
0.042395271360874176,
-0.15697690844535828,
-0.08592765033245087,
-0.039235908538103104,
0.04086203873157501,
0.02531379833817482,
-0.06479153782129288,
-0.10394936800003052,
0.04918268322944641,
-0.00038134565693326294,
-0.014792303554713726,
0.051477719098329544,
-0.08290597051382065,
0.1665632724761963,
0.15730799734592438,
0.11341086775064468,
-0.09429128468036652,
-0.05006537213921547,
-0.049747172743082047,
-0.09805992245674133,
0.05403368920087814,
-0.08081240952014923,
0.03680194169282913,
0.06566370278596878,
0.004933774936944246,
0.12807884812355042,
0.05663067847490311,
-0.046170469373464584,
-0.017491286620497704,
0.08147843182086945,
-0.1384078860282898,
-0.12795516848564148,
0.05570005625486374,
0.04672529548406601,
-0.122733473777771,
0.08926455676555634,
0.06059810146689415,
0.04172047972679138,
-0.020890476182103157,
-0.020537368953227997,
0.05566200241446495,
-0.017864352092146873,
0.14911159873008728,
0.002834389917552471,
0.07734349370002747,
-0.11976747214794159,
-0.04339396581053734,
0.04690513387322426,
-0.10856018215417862,
-0.030409187078475952,
-0.014552421867847443,
-0.13132600486278534,
-0.0962953269481659,
0.0521029569208622,
0.24990737438201904,
-0.04477507993578911,
-0.05055520310997963,
-0.06080184876918793,
-0.1867932379245758,
0.12369658052921295,
0.2589121460914612,
0.06852052360773087,
0.045230064541101456,
0.05351754277944565,
0.02034059539437294,
-0.022049367427825928,
0.08892929553985596,
0.08624447137117386,
0.07813955098390579,
-0.09775755554437637,
0.009706879034638405,
-0.04828106239438057,
0.05569169297814369,
-0.06609111279249191,
-0.007187779527157545,
-0.14306560158729553,
0.011500422842800617,
-0.11086204648017883,
0.03215761482715607,
-0.10045658051967621,
0.011767115443944931,
-0.008764765225350857,
0.00207070866599679,
-0.020186640322208405,
-0.01807151362299919,
-0.06409858912229538,
0.04433158040046692,
-0.0008824477554298937,
0.04251459240913391,
-0.09155718982219696,
0.029437217861413956,
0.02996964380145073,
-0.05625355988740921,
0.10897943377494812,
0.015071984380483627,
-0.018844079226255417,
0.03780045360326767,
-0.06447886675596237,
0.0605737529695034,
0.07698419690132141,
-0.0014755865558981895,
0.015667904168367386,
-0.006523610558360815,
0.04917243868112564,
0.0759742483496666,
0.014707457274198532,
0.08242052793502808,
0.02031070366501808,
-0.1360320746898651,
0.07767670601606369,
-0.023145519196987152,
-0.1578686386346817,
-0.049166008830070496,
0.048357293009757996,
0.06500987708568573,
-0.010370609350502491,
0.16210202872753143,
-0.080449678003788,
-0.005360613577067852,
-0.1527186632156372,
-0.009213107638061047,
0.03251528739929199,
-0.023765478283166885,
-0.04083098843693733,
-0.02439732849597931,
0.04893391951918602,
0.006306299474090338,
0.32015660405158997,
0.06343818455934525,
0.03363068401813507,
0.03970448672771454,
0.01474638469517231,
0.164746955037117,
0.04171383008360863,
0.1876714527606964,
0.01763256825506687,
-0.0567922480404377,
-0.10371902585029602,
-0.05401524528861046,
0.01636924222111702,
-0.09354973584413528,
0.08624144643545151,
0.207267627120018,
-0.08815015107393265,
0.01831624284386635,
-0.011330542154610157,
-0.0773649513721466,
-0.0437290295958519,
-0.13197384774684906,
-0.11002730578184128,
0.019644027575850487,
0.012134784832596779,
0.25700387358665466,
0.1488623470067978,
-0.04353472962975502,
0.0519656240940094,
-0.02399633452296257,
-0.08621463924646378,
-0.11456068605184555,
-0.05550732836127281,
-0.10244864225387573,
-0.16230860352516174,
0.0756150335073471,
-0.08703932166099548,
0.027327923104166985,
0.20582301914691925,
0.05261656641960144,
-0.009877319447696209,
0.10600315034389496,
-0.02741691656410694,
0.04396367073059082,
0.023426365107297897,
-0.014802979305386543,
0.0022945592645555735,
0.026263069361448288,
-0.012057582847774029,
-0.010680146515369415,
-0.006503052543848753,
0.018483351916074753,
-0.003721825312823057,
-0.016818052157759666,
0.015601332299411297,
-0.07003442198038101,
-0.022554179653525352,
0.02656346559524536,
0.013383875600993633,
-0.0325256884098053,
0.22160080075263977,
0.00009971492545446381,
0.03900982066988945,
0.015197974629700184,
0.1401098519563675,
-0.06422533094882965,
-0.0890602394938469,
-0.13022616505622864,
0.19234663248062134,
0.014428828842937946,
0.1122298464179039,
-0.00924390647560358,
-0.07261799275875092,
0.06731746345758438,
0.23509936034679413,
0.18305671215057373,
0.04865340515971184,
-0.04255151376128197,
0.0789562240242958,
-0.006342449691146612,
0.01682617887854576,
0.071770079433918,
0.009206459857523441,
0.3056102991104126,
-0.032607268542051315,
0.006295305211097002,
0.0008834745967760682,
-0.0626813992857933,
-0.0700337365269661,
0.0662890151143074,
0.04563124477863312,
-0.01409865915775299,
-0.06332620233297348,
0.08264843374490738,
-0.10687251389026642,
-0.01803886517882347,
-0.0051779053173959255,
-0.07284403592348099,
-0.0923590213060379,
-0.01355789601802826,
0.05862518399953842,
-0.02942366525530815,
0.04002552479505539,
-0.03011912852525711,
-0.012082650326192379,
0.01919371262192726,
-0.0016797258285805583,
-0.12853828072547913,
-0.08602960407733917,
0.08467648178339005,
0.01561969518661499,
0.16687226295471191,
0.008853388018906116,
0.17502886056900024,
0.10866229981184006,
-0.024279706180095673,
-0.02714875340461731,
0.12198825180530548,
0.0011711616534739733,
-0.00034970365231856704,
0.08088971674442291,
0.07618432492017746,
0.03196427598595619,
-0.005671033635735512,
0.005201298277825117,
-0.14217573404312134,
0.020874883979558945,
-0.0985819399356842,
-0.0810992568731308,
-0.1006951779127121,
0.03963422402739525,
-0.05707371607422829,
0.0809864029288292,
0.0810304805636406,
0.0005945170996710658,
0.06624744087457657,
-0.039762865751981735,
0.07164432853460312,
0.02431810460984707,
-0.006922130472958088,
-0.021093269810080528,
-0.17087222635746002,
-0.0191593449562788,
0.09246660768985748,
0.001713421894237399,
-0.30589959025382996,
-0.04189112037420273,
-0.04676803573966026,
0.0578271709382534,
-0.08976249396800995,
0.09937930107116699,
0.17006801068782806,
0.015484812669456005,
-0.047734521329402924,
-0.10699421167373657,
0.056707754731178284,
0.14246243238449097,
-0.07796919345855713,
-0.03593267500400543
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# emotion_classification
This model is a fine-tuned version of [dennisjooo/emotion_classification](https://huggingface.co/dennisjooo/emotion_classification) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7891
- Accuracy: 0.7575
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine_with_restarts
- num_epochs: 10
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.7123 | 1.0 | 25 | 0.8681 | 0.735 |
| 0.6349 | 2.0 | 50 | 0.8721 | 0.73 |
| 0.6354 | 3.0 | 75 | 0.8732 | 0.725 |
| 0.6189 | 4.0 | 100 | 0.8406 | 0.735 |
| 0.6364 | 5.0 | 125 | 0.8456 | 0.74 |
| 0.5833 | 6.0 | 150 | 0.8503 | 0.725 |
| 0.5384 | 7.0 | 175 | 0.8023 | 0.755 |
| 0.5297 | 8.0 | 200 | 0.8002 | 0.7525 |
| 0.5487 | 9.0 | 225 | 0.8253 | 0.745 |
| 0.5068 | 10.0 | 250 | 0.7891 | 0.7575 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "dennisjooo/emotion_classification", "model-index": [{"name": "emotion_classification", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.7575, "name": "Accuracy"}]}]}]} | image-classification | mhdiqbalpradipta/emotion_classification | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:dennisjooo/emotion_classification",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T07:46:12+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-dennisjooo/emotion_classification #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| emotion\_classification
=======================
This model is a fine-tuned version of dennisjooo/emotion\_classification on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7891
* Accuracy: 0.7575
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine\_with\_restarts
* num\_epochs: 10
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\\_with\\_restarts\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-dennisjooo/emotion_classification #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\\_with\\_restarts\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
82,
122,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-dennisjooo/emotion_classification #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\\_with\\_restarts\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.1424366533756256,
0.16143907606601715,
-0.002839229768142104,
0.1075059175491333,
0.1452186107635498,
0.041060253977775574,
0.13361984491348267,
0.11955316364765167,
-0.060690443962812424,
0.082107774913311,
0.1035030335187912,
0.07971885800361633,
0.06623577326536179,
0.15925724804401398,
-0.061091531068086624,
-0.2507229447364807,
0.04219932109117508,
0.04174472391605377,
-0.04276818782091141,
0.12554489076137543,
0.09500826150178909,
-0.12793220579624176,
0.09829726070165634,
0.020789993926882744,
-0.17130817472934723,
-0.04722233861684799,
0.004419357981532812,
-0.05088343098759651,
0.11549453437328339,
0.0148134371265769,
0.12417196482419968,
0.05882770195603371,
0.09640338271856308,
-0.17438410222530365,
0.021873200312256813,
0.059566665440797806,
-0.0013224307913333178,
0.09643957018852234,
0.07958003878593445,
0.008389274589717388,
0.058990493416786194,
-0.08225297182798386,
0.06145351752638817,
0.015060522593557835,
-0.10513455420732498,
-0.20503078401088715,
-0.09258971363306046,
0.10005895793437958,
0.09985482692718506,
0.08006379753351212,
-0.00886580441147089,
0.1214505285024643,
-0.06145443767309189,
0.08337624371051788,
0.21588851511478424,
-0.2231377512216568,
-0.08405352383852005,
-0.0070573920384049416,
0.03366586193442345,
0.04242396354675293,
-0.11181054264307022,
-0.039627011865377426,
0.03763222694396973,
0.027099071070551872,
0.11827372014522552,
0.004499218892306089,
0.00036620075115934014,
-0.03802504390478134,
-0.13092385232448578,
-0.08976876735687256,
0.1352485567331314,
0.08978389948606491,
-0.042186181992292404,
-0.06834232062101364,
-0.06988433003425598,
-0.1922103464603424,
-0.03885757550597191,
0.009565223939716816,
0.03039950504899025,
-0.045686572790145874,
-0.07921477407217026,
-0.0015611263224855065,
-0.12115903198719025,
-0.04966919496655464,
0.009070388972759247,
0.12013039737939835,
0.04075208678841591,
0.017024142667651176,
0.008038542233407497,
0.12049586325883865,
0.03640288859605789,
-0.16277417540550232,
0.00004314192847232334,
-0.00012854422675445676,
-0.0006769869942218065,
-0.018999304622411728,
-0.03666019067168236,
-0.0314207449555397,
0.007688250858336687,
0.11226836591959,
-0.07123105973005295,
0.026517048478126526,
0.0050827255472540855,
0.041430842131376266,
-0.07603903114795685,
0.1817346066236496,
-0.0624057874083519,
-0.0005894046043977141,
0.02559013105928898,
0.15670080482959747,
0.04323180764913559,
-0.013669914565980434,
-0.10453726351261139,
0.015370909124612808,
0.15159527957439423,
0.021260879933834076,
-0.04111037403345108,
0.05497761815786362,
-0.08065155893564224,
-0.03871724009513855,
0.10519962012767792,
-0.07345534861087799,
0.024073205888271332,
0.03280043601989746,
-0.06483247131109238,
-0.015440392307937145,
0.03209160640835762,
0.008527115918695927,
-0.0029679248109459877,
0.08462315797805786,
-0.08833345025777817,
-0.0059365383349359035,
-0.08008550107479095,
-0.09648867696523666,
0.04061634838581085,
-0.08461349457502365,
0.013276739977300167,
-0.11867128312587738,
-0.15124855935573578,
-0.03134601563215256,
0.040403660386800766,
-0.01692776568233967,
-0.08303551375865936,
-0.04174242541193962,
-0.09756046533584595,
0.03303750976920128,
-0.005937531124800444,
0.07064398378133774,
-0.05013173073530197,
0.11278746277093887,
0.025539711117744446,
0.062458109110593796,
-0.008768463507294655,
0.040839750319719315,
-0.10023446381092072,
0.056624870747327805,
-0.1435467004776001,
0.04023881256580353,
-0.061103083193302155,
0.059449173510074615,
-0.0798717513680458,
-0.11540821939706802,
0.04202694445848465,
-0.04746876657009125,
0.07466385513544083,
0.11847222596406937,
-0.14440059661865234,
-0.07106605917215347,
0.1564820557832718,
-0.11834130436182022,
-0.1215493232011795,
0.12836439907550812,
-0.026653394103050232,
-0.02667928673326969,
0.032483261078596115,
0.13238689303398132,
0.08338971436023712,
-0.09594900161027908,
-0.04021868854761124,
-0.03600502386689186,
0.08233045041561127,
-0.01783565990626812,
0.10968714952468872,
0.06507608294487,
0.0008315778104588389,
0.02273394726216793,
-0.10387023538351059,
0.08562992513179779,
-0.09729518741369247,
-0.09810972213745117,
-0.03830432519316673,
-0.09126061946153641,
0.0648130476474762,
0.07870714366436005,
0.02714960277080536,
-0.08212026208639145,
-0.10683413594961166,
-0.0007265538442879915,
0.12541687488555908,
-0.07909950613975525,
0.004675623960793018,
-0.0821206271648407,
0.1254395693540573,
-0.06831033527851105,
-0.026413865387439728,
-0.15555340051651,
-0.06737867742776871,
0.03407149389386177,
-0.02881883829832077,
-0.01744811050593853,
-0.043539147824048996,
0.05196978151798248,
0.09039703756570816,
-0.04916892573237419,
-0.09562329202890396,
-0.10111788660287857,
-0.0028139350470155478,
-0.07746046781539917,
-0.21278533339500427,
-0.07562796771526337,
-0.02962009608745575,
0.2305229753255844,
-0.20982573926448822,
0.03186962008476257,
0.02088022604584694,
0.12665992975234985,
0.04378272593021393,
-0.033595163375139236,
-0.009988994337618351,
0.05073685571551323,
-0.04205269366502762,
-0.06930308789014816,
0.06505018472671509,
0.033321600407361984,
-0.06645539402961731,
0.012160495854914188,
-0.13048765063285828,
0.12256252765655518,
0.12214282155036926,
-0.008345309644937515,
-0.0694110319018364,
-0.02026107907295227,
-0.07621849328279495,
-0.026220442727208138,
-0.027677271515130997,
0.02126331999897957,
0.09445206075906754,
0.021153567358851433,
0.1405440717935562,
-0.08856429904699326,
-0.006825758144259453,
0.05267021059989929,
-0.01859927922487259,
-0.02879551611840725,
0.10095635056495667,
0.09122608602046967,
-0.12257219851016998,
0.14877164363861084,
0.13687320053577423,
-0.08071232587099075,
0.11151960492134094,
-0.051605310291051865,
-0.07570402324199677,
-0.03690929338335991,
-0.0014189801877364516,
0.01688467711210251,
0.1469557285308838,
-0.09843702614307404,
-0.008493582718074322,
0.03336653113365173,
-0.032086990773677826,
-0.007677057292312384,
-0.22911469638347626,
-0.022886930033564568,
0.033579498529434204,
-0.037257276475429535,
0.0074924565851688385,
-0.005992954596877098,
0.020638197660446167,
0.10271243005990982,
0.00770933972671628,
-0.07804333418607712,
0.0158877894282341,
-0.003594739828258753,
-0.07466906309127808,
0.18505078554153442,
-0.07962927967309952,
-0.20871227979660034,
-0.12784412503242493,
0.00906453188508749,
-0.06895704567432404,
-0.0017235134728252888,
0.044531453400850296,
-0.07556699216365814,
-0.03699318692088127,
-0.08855642378330231,
-0.03501230105757713,
0.03180951625108719,
0.018074670806527138,
0.033260688185691833,
-0.003148954128846526,
0.06597556173801422,
-0.07636310905218124,
-0.004729113541543484,
-0.013132265768945217,
-0.03840811178088188,
0.06972162425518036,
0.027977177873253822,
0.1192217767238617,
0.1431220918893814,
-0.018198076635599136,
0.008089488372206688,
-0.03288046270608902,
0.21981507539749146,
-0.08978907018899918,
-0.01069861464202404,
0.15274126827716827,
-0.0351669006049633,
0.07781513035297394,
0.14389358460903168,
0.04159129038453102,
-0.06859192997217178,
0.000028376349291647784,
0.011032185517251492,
-0.02152743563055992,
-0.18573318421840668,
-0.042993053793907166,
-0.03711989149451256,
0.01073555275797844,
0.11572230607271194,
0.029546694830060005,
0.04733008146286011,
0.08528289198875427,
-0.0309736430644989,
0.027713773772120476,
-0.018601074814796448,
0.08924634009599686,
0.08782926946878433,
0.03804296255111694,
0.1028294563293457,
-0.014878963120281696,
-0.03871363028883934,
0.043248459696769714,
0.010536490008234978,
0.21184134483337402,
-0.04292202368378639,
0.1413378268480301,
0.03498818725347519,
0.20183901488780975,
0.019289372488856316,
0.06645989418029785,
0.0003507425426505506,
-0.0038959102239459753,
-0.011774664744734764,
-0.039210036396980286,
-0.0633883997797966,
0.025842951610684395,
-0.01246584951877594,
0.07629958540201187,
-0.14477889239788055,
0.021690476685762405,
0.026311151683330536,
0.311061829328537,
0.05391441285610199,
-0.3798767328262329,
-0.11290494352579117,
-0.00045708465040661395,
-0.017730168998241425,
-0.050710562616586685,
-0.01599840633571148,
0.11680787056684494,
-0.09490104764699936,
0.05261914059519768,
-0.09442180395126343,
0.07758333534002304,
-0.07454507052898407,
0.013121306896209717,
0.06926073879003525,
0.06426022201776505,
0.022844599559903145,
0.06359480321407318,
-0.19955535233020782,
0.26257026195526123,
-0.014712312258780003,
0.04382164031267166,
-0.059657543897628784,
0.0031376155093312263,
0.06714237481355667,
0.08279673755168915,
0.07845888286828995,
-0.004483568947762251,
-0.015401597134768963,
-0.22542104125022888,
-0.12164898216724396,
0.014112187549471855,
0.06932003051042557,
-0.05391604080796242,
0.11287401616573334,
-0.031987253576517105,
-0.025609884411096573,
0.038588181138038635,
0.02280372381210327,
-0.07941364496946335,
-0.09132357686758041,
0.030941041186451912,
0.022296182811260223,
0.03375225514173508,
-0.09897597134113312,
-0.1351119726896286,
-0.0664030909538269,
0.12831059098243713,
-0.016734419390559196,
-0.04519173130393028,
-0.12748199701309204,
0.09500891715288162,
0.0892183855175972,
-0.09204868227243423,
0.04075126349925995,
-0.02319580502808094,
0.1784289926290512,
0.011285839602351189,
-0.06699319183826447,
0.06919647008180618,
-0.07267117500305176,
-0.2323140799999237,
-0.04605202004313469,
0.12152176350355148,
0.018312664702534676,
0.05611436069011688,
0.007302406709641218,
0.01763136126101017,
-0.02197897993028164,
-0.05739423632621765,
0.03414813429117203,
0.03993271663784981,
0.03995724767446518,
0.019987286999821663,
-0.009893662296235561,
-0.031088976189494133,
-0.073258176445961,
-0.012803180143237114,
0.13936595618724823,
0.2624311149120331,
-0.08346755057573318,
0.003493252908810973,
0.026279186829924583,
-0.04381374642252922,
-0.1895618587732315,
0.021098770201206207,
0.11536210775375366,
0.02319309487938881,
-0.0015109814703464508,
-0.14771240949630737,
0.054799728095531464,
0.07788977771997452,
-0.04463810473680496,
0.09085449576377869,
-0.2397812157869339,
-0.11531148850917816,
0.09436198323965073,
0.1563255488872528,
0.07609573751688004,
-0.1585685759782791,
-0.028368333354592323,
-0.007591659668833017,
-0.12480492889881134,
0.15385045111179352,
-0.03315787389874458,
0.09926926344633102,
-0.016820671036839485,
0.0677536353468895,
0.022758103907108307,
-0.054814573377370834,
0.1477692872285843,
-0.0360141396522522,
0.07286422699689865,
-0.04908221960067749,
-0.031357381492853165,
0.0696786493062973,
-0.07960242033004761,
0.02636696584522724,
-0.04638708755373955,
0.049899425357580185,
-0.10259265452623367,
-0.01165876816958189,
-0.11272121965885162,
0.019665861502289772,
-0.03686458617448807,
-0.02940157987177372,
-0.03083803690969944,
0.06809700280427933,
0.06668016314506531,
0.004237378481775522,
0.1278863400220871,
0.04001742601394653,
0.1141197457909584,
0.1215766966342926,
0.08013685047626495,
-0.02881445549428463,
-0.06099887564778328,
-0.05006099492311478,
-0.032616499811410904,
0.07116968184709549,
-0.1175234392285347,
0.024701809510588646,
0.12403993308544159,
0.026106594130396843,
0.14800597727298737,
0.05934799090027809,
-0.057395439594984055,
0.026959344744682312,
0.08113979548215866,
-0.14174163341522217,
-0.10362179577350616,
-0.04514997825026512,
0.03507138788700104,
-0.15053999423980713,
0.009798520244657993,
0.11179330199956894,
-0.0898427665233612,
-0.017360761761665344,
-0.00387369841337204,
0.02241704799234867,
0.0010031720157712698,
0.14292208850383759,
0.06334695965051651,
0.05099272355437279,
-0.10407432168722153,
0.07494749873876572,
0.05940227955579758,
-0.16095460951328278,
0.0072427052073180676,
0.05315060168504715,
-0.09369729459285736,
-0.04486342519521713,
0.037886470556259155,
0.13700184226036072,
-0.06519906222820282,
-0.06181531772017479,
-0.13798397779464722,
-0.1215059757232666,
0.0641036331653595,
0.13256098330020905,
0.07836651057004929,
0.026069555431604385,
-0.0076853143982589245,
0.024979811161756516,
-0.12130816280841827,
0.1194954589009285,
0.06490109115839005,
0.09405682981014252,
-0.17737427353858948,
0.1046488806605339,
0.007544289343059063,
0.05864753946661949,
-0.018358128145337105,
0.018185216933488846,
-0.07700515538454056,
-0.015857290476560593,
-0.11430635303258896,
0.03500216081738472,
-0.04575355350971222,
0.009359081275761127,
0.0014420836232602596,
-0.05596575886011124,
-0.04455031827092171,
0.01709880493581295,
-0.10406558215618134,
-0.04088353365659714,
0.010529184713959694,
0.0530698262155056,
-0.10605642944574356,
-0.0373581163585186,
0.03481375426054001,
-0.10593003779649734,
0.1015404611825943,
0.030066393315792084,
0.024525048211216927,
0.029872896149754524,
-0.08573875576257706,
-0.008407535962760448,
0.07296670228242874,
0.011854241602122784,
0.03638102114200592,
-0.10357696563005447,
0.01190469041466713,
-0.012804483994841576,
-0.015104076825082302,
-0.005387941841036081,
0.10573779046535492,
-0.11457443982362747,
-0.013250933960080147,
0.019059056416153908,
-0.03344118222594261,
-0.06578802317380905,
0.04343697801232338,
0.06249568983912468,
0.015616754069924355,
0.18781113624572754,
-0.07011053711175919,
0.04434845596551895,
-0.23002897202968597,
-0.0026592370122671127,
-0.02283635176718235,
-0.09969072043895721,
-0.12418092787265778,
-0.032905105501413345,
0.07198681682348251,
-0.05333483964204788,
0.06217123195528984,
0.0037986671086400747,
0.07311264425516129,
0.015365981496870518,
0.021830959245562553,
0.015226907096803188,
0.04338543117046356,
0.15874923765659332,
0.03549222648143768,
-0.03575422242283821,
0.06773030012845993,
0.0019530483987182379,
0.08259659260511398,
0.05798386037349701,
0.16262955963611603,
0.1296369582414627,
0.04388445243239403,
0.08906817436218262,
0.06958737224340439,
-0.024698078632354736,
-0.15961818397045135,
0.020223041996359825,
-0.0742151066660881,
0.12270941585302353,
-0.00022167421411722898,
0.2271239161491394,
0.09939414262771606,
-0.17155802249908447,
0.03528508543968201,
-0.04258756339550018,
-0.05826672911643982,
-0.06694146245718002,
-0.10318992286920547,
-0.09755540639162064,
-0.1443421095609665,
0.013989393599331379,
-0.1242300271987915,
-0.008590986020863056,
0.07999758422374725,
0.0037770147901028395,
-0.02231847308576107,
0.11638248711824417,
0.05063342675566673,
0.011567307636141777,
0.10113117843866348,
0.031085867434740067,
-0.038564518094062805,
-0.049777109175920486,
-0.09102044999599457,
0.03951157629489899,
0.011367172934114933,
0.06434948742389679,
-0.05346401035785675,
-0.04074670746922493,
0.06439679861068726,
0.014599024318158627,
-0.12650351226329803,
0.028384974226355553,
-0.014460557140409946,
0.054572295397520065,
0.03848348185420036,
-0.016035793349146843,
0.050221092998981476,
-0.005444154608994722,
0.18061058223247528,
-0.058755069971084595,
-0.0166364386677742,
-0.13018569350242615,
0.16842423379421234,
-0.029560238122940063,
-0.0500003919005394,
0.062374528497457504,
-0.08733298629522324,
-0.006194163113832474,
0.17544879019260406,
0.14095833897590637,
-0.06889338791370392,
-0.002590074436739087,
0.007147269789129496,
-0.007140934932976961,
-0.0757497027516365,
0.10354658961296082,
0.11035893112421036,
0.016526231542229652,
-0.09823266416788101,
-0.027550190687179565,
-0.061014268547296524,
-0.014632060192525387,
-0.010597839951515198,
0.022620167583227158,
0.013792973011732101,
0.003043064149096608,
-0.07093537598848343,
0.052333928644657135,
-0.03446754068136215,
-0.07746995985507965,
0.0864734947681427,
-0.2053004801273346,
-0.1793731451034546,
-0.035223908722400665,
0.09221332520246506,
0.02216697670519352,
0.042045991867780685,
-0.00996970571577549,
0.0015699658542871475,
0.12189318239688873,
-0.027560723945498466,
-0.06168342009186745,
-0.0737556517124176,
0.07804959267377853,
-0.07828132063150406,
0.23079371452331543,
-0.04406019300222397,
0.03648217022418976,
0.1095890924334526,
0.04510515183210373,
-0.11776537448167801,
0.028220398351550102,
0.06808960437774658,
-0.07646819949150085,
0.02687603235244751,
0.17472036182880402,
-0.03183767944574356,
0.08592812716960907,
0.04108725115656853,
-0.15106014907360077,
-0.010117274709045887,
-0.04208892583847046,
-0.0390453040599823,
-0.04035373404622078,
0.008006570860743523,
-0.05126023292541504,
0.12901873886585236,
0.18980857729911804,
-0.047281663864851,
-0.03201010450720787,
-0.05749514326453209,
0.00988743081688881,
0.07423514872789383,
0.06116674467921257,
-0.012320371344685555,
-0.21422170102596283,
0.010639077983796597,
-0.00727889034897089,
0.02154999040067196,
-0.23838220536708832,
-0.09067864716053009,
-0.009432870894670486,
-0.05892525613307953,
-0.10300874710083008,
0.09466417133808136,
0.05571696162223816,
0.0412425771355629,
-0.04325634241104126,
-0.025379745289683342,
-0.06454388052225113,
0.16170619428157806,
-0.14143525063991547,
-0.07893183082342148
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# OndeviceAI-base
This model is a fine-tuned version of [paust/pko-t5-base](https://huggingface.co/paust/pko-t5-base) on the None dataset.
## How to use
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
from typing import List
tokenizer = AutoTokenizer.from_pretrained("yeye776/OndeviceAI-base")
model = AutoModelForSeq2SeqLM.from_pretrained("yeye776/OndeviceAI-base")
prompt = "분류 및 인식해줘 :"
def prepare_input(question: str):
inputs = f"{prompt} {question}"
input_ids = tokenizer(inputs, max_length=700, return_tensors="pt").input_ids
return input_ids
def inference(question: str) -> str:
input_data = prepare_input(question=question)
input_data = input_data.to(model.device)
outputs = model.generate(inputs=input_data, num_beams=10, top_k=10, max_length=1024)
result = tokenizer.decode(token_ids=outputs[0], skip_special_tokens=True)
return result
inference("안방 조명 켜줘")
```
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0007
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.06
- num_epochs: 10
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "cc-by-4.0", "tags": ["generated_from_trainer"], "base_model": "paust/pko-t5-base", "model-index": [{"name": "OndeviceAI-base", "results": []}]} | text2text-generation | yeye776/OndeviceAI-base | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:paust/pko-t5-base",
"license:cc-by-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T08:03:28+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# OndeviceAI-base
This model is a fine-tuned version of paust/pko-t5-base on the None dataset.
## How to use
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0007
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.06
- num_epochs: 10
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# OndeviceAI-base\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.",
"## How to use",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 10",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# OndeviceAI-base\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.",
"## How to use",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 10",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
83,
33,
4,
129,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-paust/pko-t5-base #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# OndeviceAI-base\n\nThis model is a fine-tuned version of paust/pko-t5-base on the None dataset.## How to use### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0007\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.06\n- num_epochs: 10### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.11100435256958008,
0.12467678636312485,
-0.003916221205145121,
0.09340550005435944,
0.10593501478433609,
0.04339111968874931,
0.15257678925991058,
0.13470271229743958,
-0.0571369007229805,
0.11118064820766449,
0.09331067651510239,
0.05356017127633095,
0.06066760793328285,
0.15487125515937805,
-0.04272353649139404,
-0.19019551575183868,
0.025309376418590546,
-0.04900951683521271,
-0.031025469303131104,
0.10300090909004211,
0.08301004767417908,
-0.09864816069602966,
0.06926967203617096,
-0.00708976574242115,
-0.1368670016527176,
-0.0441756397485733,
-0.02951812371611595,
-0.04573310911655426,
0.09049572050571442,
0.005087961442768574,
0.1185031458735466,
0.04108961299061775,
0.1082935631275177,
-0.1860588937997818,
0.0026685548946261406,
0.07819616794586182,
0.017892085015773773,
0.09935083985328674,
0.09618574380874634,
-0.005181396380066872,
0.04042346030473709,
-0.11383488774299622,
0.07740449160337448,
0.024613864719867706,
-0.1063648909330368,
-0.023338261991739273,
-0.10825291275978088,
0.07101286947727203,
0.10496565699577332,
0.09449870884418488,
0.00036827928852289915,
0.10413186997175217,
-0.08664936572313309,
0.06691093742847443,
0.17237161099910736,
-0.2566242218017578,
-0.06030750274658203,
0.055558495223522186,
0.03586055338382721,
0.08561258018016815,
-0.10007847100496292,
-0.04443274438381195,
0.020906012505292892,
0.014740798622369766,
0.07106205821037292,
-0.013501672074198723,
-0.05093848705291748,
-0.018895482644438744,
-0.09996535629034042,
-0.05327260494232178,
0.10801846534013748,
0.03616257756948471,
-0.03623073548078537,
-0.08935034275054932,
-0.0579381138086319,
-0.16440677642822266,
-0.023599909618496895,
0.01260468177497387,
0.021297737956047058,
-0.021421749144792557,
-0.05069253221154213,
-0.028954220935702324,
-0.08326894789934158,
-0.061371102929115295,
-0.000089013330580201,
0.06623811274766922,
0.026462281122803688,
0.0038776365108788013,
-0.0023038885556161404,
0.11987742781639099,
0.01331128180027008,
-0.1307523399591446,
0.00141045986674726,
-0.006954776123166084,
-0.09509968757629395,
-0.05821891501545906,
-0.03275956213474274,
-0.004999714903533459,
0.03327649459242821,
0.1564285159111023,
-0.029316505417227745,
0.06436875462532043,
0.000560765853151679,
0.016499143093824387,
-0.038724206387996674,
0.14742767810821533,
-0.03998933732509613,
-0.046182163059711456,
0.030797570943832397,
0.10446355491876602,
0.014887291938066483,
-0.02420457825064659,
-0.08533678203821182,
-0.022146815434098244,
0.09607687592506409,
0.061710841953754425,
-0.009038642048835754,
0.029189394786953926,
-0.055192239582538605,
-0.022316643968224525,
0.07687310129404068,
-0.11135901510715485,
0.044085945934057236,
0.03355244919657707,
-0.036336954683065414,
-0.0027096113190054893,
0.019235536456108093,
0.003529280424118042,
-0.05640676990151405,
0.0825672373175621,
-0.08791667968034744,
-0.019024772569537163,
-0.05878259614109993,
-0.06732819974422455,
0.03312550485134125,
-0.11571934819221497,
-0.040472716093063354,
-0.08751164376735687,
-0.15273568034172058,
-0.05652862787246704,
0.020104525610804558,
-0.05958553031086922,
-0.06191130355000496,
-0.06230534613132477,
-0.08999243378639221,
0.04426192864775658,
-0.004422422964125872,
0.10670057684183121,
-0.056326620280742645,
0.06008613854646683,
-0.05876933038234711,
0.020152544602751732,
0.0593821220099926,
0.0339813232421875,
-0.07214375585317612,
0.04318656399846077,
-0.1426004320383072,
0.07490061968564987,
-0.06391708552837372,
0.03261391818523407,
-0.11117466539144516,
-0.056005410850048065,
0.02209389954805374,
-0.04366243630647659,
0.06520316004753113,
0.14530175924301147,
-0.16864271461963654,
-0.010606091469526291,
0.15072984993457794,
-0.06384041905403137,
-0.07552196085453033,
0.09716808795928955,
-0.0398368164896965,
-0.03709760308265686,
0.043629806488752365,
0.13032831251621246,
0.10738520324230194,
-0.12162593752145767,
-0.017674081027507782,
0.004026063717901707,
0.07930559664964676,
0.008253834210336208,
0.07065316289663315,
-0.0329505130648613,
0.015909774228930473,
0.005779980216175318,
-0.03994588926434517,
0.010010397993028164,
-0.05636141076683998,
-0.07124832272529602,
-0.056029312312603,
-0.08246637880802155,
0.00922621414065361,
0.003921015188097954,
0.031486108899116516,
-0.06158720701932907,
-0.11325425654649734,
0.0009673216263763607,
0.11319085955619812,
-0.08154980838298798,
0.02513393759727478,
-0.0653965100646019,
0.09673389047384262,
-0.041124992072582245,
-0.027213629335165024,
-0.1746031641960144,
-0.1254185438156128,
0.050168756395578384,
-0.015928536653518677,
0.0236094668507576,
-0.017515383660793304,
0.047160133719444275,
0.075541190803051,
-0.03334597125649452,
-0.024596285074949265,
-0.03195466473698616,
-0.004597740713506937,
-0.09443778544664383,
-0.19902926683425903,
-0.030050108209252357,
-0.027008110657334328,
0.20986108481884003,
-0.19373291730880737,
0.002682592486962676,
-0.014397301711142063,
0.1495412141084671,
0.02521989494562149,
-0.05170468986034393,
0.005892196670174599,
0.05918523296713829,
-0.02149018459022045,
-0.09849372506141663,
0.035637788474559784,
0.022802330553531647,
-0.08083028346300125,
0.007339184638112783,
-0.16790734231472015,
-0.00758156506344676,
0.09709614515304565,
0.045055415481328964,
-0.10670939087867737,
-0.057225219905376434,
-0.04677141457796097,
-0.0373711884021759,
-0.04622286185622215,
0.0016983702080324292,
0.15824119746685028,
0.03248969838023186,
0.1305731236934662,
-0.07564400881528854,
-0.06368574500083923,
0.013839859515428543,
0.005881917662918568,
0.01830494962632656,
0.09779369831085205,
0.09009779989719391,
-0.07363487035036087,
0.09354878962039948,
0.08388704806566238,
-0.06514336913824081,
0.1009642705321312,
-0.0617852546274662,
-0.09150181710720062,
-0.02458338998258114,
0.03403076156973839,
0.025588948279619217,
0.08793002367019653,
-0.05643214285373688,
0.020153816789388657,
0.044369518756866455,
0.013191203586757183,
0.021827135235071182,
-0.1688442826271057,
0.000691175286192447,
0.007994815707206726,
-0.050817329436540604,
-0.013300949707627296,
-0.03510205075144768,
0.036330774426460266,
0.09248726814985275,
0.007309895474463701,
-0.04383763670921326,
0.0126182297244668,
-0.017782164737582207,
-0.09021425992250443,
0.18477216362953186,
-0.11277255415916443,
-0.10762611031532288,
-0.11591368168592453,
0.016073936596512794,
-0.046078309416770935,
-0.0021610793191939592,
-0.00133990787435323,
-0.04146676883101463,
-0.052643343806266785,
-0.11132945120334625,
-0.059417352080345154,
0.006175375077873468,
0.0076342797838151455,
-0.0004412643611431122,
0.010148079134523869,
0.040896277874708176,
-0.09435920417308807,
-0.004743786994367838,
0.012962998822331429,
-0.058517202734947205,
0.041339483112096786,
0.018182313069701195,
0.07111847400665283,
0.1621980220079422,
-0.0022342056035995483,
-0.005986540578305721,
-0.034588538110256195,
0.20689833164215088,
-0.09407985955476761,
0.03160950541496277,
0.09009063243865967,
-0.01965726725757122,
0.06228527054190636,
0.13576926290988922,
0.017396220937371254,
-0.05992672219872475,
0.010881498456001282,
0.035951852798461914,
-0.017694922164082527,
-0.2559511959552765,
-0.04055919870734215,
-0.05978796258568764,
-0.003383210627362132,
0.12034863978624344,
0.03897257149219513,
0.03380393236875534,
0.0642821416258812,
-0.01803298108279705,
0.04264016076922417,
0.03193412721157074,
0.09028257429599762,
0.06739004701375961,
0.07338084280490875,
0.10896174609661102,
-0.029682226479053497,
-0.007919993251562119,
0.06729280948638916,
0.045434944331645966,
0.212349072098732,
-0.02184942178428173,
0.18170350790023804,
0.009549950249493122,
0.13842856884002686,
0.00954463705420494,
0.0627841055393219,
0.021982965990900993,
0.01134718768298626,
0.006330778356641531,
-0.06467005610466003,
-0.021953053772449493,
0.03420468419790268,
0.021387865766882896,
-0.012260008603334427,
-0.0686032697558403,
0.0305330827832222,
0.031556934118270874,
0.25951826572418213,
0.07283125072717667,
-0.30484849214553833,
-0.04659849405288696,
0.01894369162619114,
-0.023130938410758972,
-0.056913360953330994,
0.014248455874621868,
0.10095158964395523,
-0.12245673686265945,
0.046245694160461426,
-0.051637228578329086,
0.09718523919582367,
-0.08192574977874756,
-0.01497103925794363,
0.048949744552373886,
0.09743259847164154,
-0.013292311690747738,
0.06462890654802322,
-0.19001537561416626,
0.18335676193237305,
0.011249080300331116,
0.07237690687179565,
-0.0701424777507782,
0.033122166991233826,
0.01484769582748413,
0.05039439722895622,
0.09807561337947845,
-0.0013301981380209327,
-0.058344729244709015,
-0.12682917714118958,
-0.14169661700725555,
0.022251354530453682,
0.09344535320997238,
-0.055519115179777145,
0.08630304783582687,
-0.044900692999362946,
-0.01190536841750145,
0.027561413124203682,
-0.05052332580089569,
-0.10661112517118454,
-0.13619089126586914,
0.03179791569709778,
0.015144026838243008,
-0.013270611874759197,
-0.05562209337949753,
-0.0831063836812973,
-0.07469920068979263,
0.18315738439559937,
-0.05142923817038536,
-0.049994248896837234,
-0.143140971660614,
0.001826285501010716,
0.14090554416179657,
-0.053390081971883774,
0.027974294498562813,
-0.016112592071294785,
0.13045834004878998,
0.0024986923672258854,
-0.06200643256306648,
0.060048818588256836,
-0.06053335964679718,
-0.2146332859992981,
-0.07272530347108841,
0.13399268686771393,
0.032445259392261505,
0.045555345714092255,
0.00438426248729229,
0.031112512573599815,
-0.01437010895460844,
-0.08875623345375061,
0.007728650234639645,
0.07679029554128647,
0.10629474371671677,
0.016310621052980423,
-0.04939023032784462,
0.02441985346376896,
-0.03846137598156929,
-0.007199447136372328,
0.11940329521894455,
0.2791231870651245,
-0.07859274744987488,
0.0730157271027565,
0.07313862442970276,
-0.06322725117206573,
-0.1574256271123886,
0.005973807070404291,
0.08871585875749588,
0.03047039359807968,
-0.02915998548269272,
-0.15168620645999908,
0.08757075667381287,
0.10862313210964203,
-0.015279214829206467,
0.0474352203309536,
-0.29146623611450195,
-0.14673489332199097,
0.0819564163684845,
0.07390523701906204,
-0.004883831366896629,
-0.12819026410579681,
-0.04524591192603111,
-0.033556658774614334,
-0.10692635178565979,
0.1278638243675232,
-0.0688457041978836,
0.11700880527496338,
-0.005587565712630749,
0.05671229586005211,
0.027894310653209686,
-0.04749326407909393,
0.15214058756828308,
0.03855873644351959,
0.04747132956981659,
-0.04145999252796173,
0.005044654943048954,
0.06128477305173874,
-0.08420459926128387,
0.014618007466197014,
-0.07696497440338135,
0.06447255611419678,
-0.1293276995420456,
-0.01671599969267845,
-0.05369759723544121,
0.02766351029276848,
-0.04959587752819061,
-0.044871799647808075,
-0.04355856031179428,
0.04374510049819946,
0.0769856795668602,
-0.021584948524832726,
0.07186365127563477,
-0.0001173712735180743,
0.07824647426605225,
0.13762332499027252,
0.08762872964143753,
0.007592698093503714,
-0.05931565538048744,
0.0030320780351758003,
-0.0382644422352314,
0.027506979182362556,
-0.09524402767419815,
0.017480449751019478,
0.12568213045597076,
0.027990248054265976,
0.11898461729288101,
0.048192303627729416,
-0.05114646255970001,
-0.0042668175883591175,
0.04112398624420166,
-0.13416817784309387,
-0.04881148040294647,
0.01752212457358837,
-0.02839241735637188,
-0.1172855794429779,
-0.03514315187931061,
0.14689835906028748,
-0.03624739497900009,
-0.0166108887642622,
0.0093008354306221,
0.018433213233947754,
-0.016624391078948975,
0.16198621690273285,
0.043029241263866425,
0.06475086510181427,
-0.08844789117574692,
0.11885130405426025,
0.0941183865070343,
-0.07200461626052856,
0.023759597912430763,
0.0974607765674591,
-0.11132337152957916,
-0.01700924150645733,
0.07878737896680832,
0.1421729177236557,
-0.04694831743836403,
-0.03239207714796066,
-0.06802500784397125,
-0.10350409895181656,
0.07675148546695709,
0.12041519582271576,
0.02050815522670746,
0.02688496559858322,
-0.026842167600989342,
0.0010472771245986223,
-0.13503454625606537,
0.09215608239173889,
0.05344517529010773,
0.08171579241752625,
-0.10298830270767212,
0.12187524884939194,
-0.020201215520501137,
0.009579258039593697,
-0.0063405875116586685,
0.0335981585085392,
-0.12262442708015442,
-0.03550639748573303,
-0.11989182978868484,
0.04874488338828087,
-0.04277994856238365,
0.0007601386751048267,
-0.002913752803578973,
-0.020715102553367615,
-0.03181295096874237,
0.009504351764917374,
-0.0647365152835846,
-0.057292256504297256,
-0.020685972645878792,
0.05537541210651398,
-0.13015373051166534,
-0.003444226924329996,
0.02390839345753193,
-0.11438561975955963,
0.09648013114929199,
0.046985432505607605,
0.011745437979698181,
-0.0021836566738784313,
-0.12321045994758606,
0.017406579107046127,
0.01134686078876257,
0.024226779118180275,
0.05262075364589691,
-0.107358917593956,
-0.001036005443893373,
-0.02918931096792221,
-0.01035621389746666,
0.020208949223160744,
0.05434870347380638,
-0.1047983318567276,
0.003233938245102763,
-0.02071695774793625,
-0.0649200901389122,
-0.07260840386152267,
0.043181050568819046,
0.11368639022111893,
0.0031142057850956917,
0.12686705589294434,
-0.06140074506402016,
0.0679325982928276,
-0.18387092649936676,
-0.027142515406012535,
0.01388649269938469,
-0.03515082225203514,
-0.0630873367190361,
-0.050149425864219666,
0.09419215470552444,
-0.04491231217980385,
0.0965137779712677,
-0.03256833553314209,
0.058423083275556564,
0.01596478931605816,
-0.0063709113746881485,
-0.0003071733517572284,
0.05717635899782181,
0.16219022870063782,
0.05339228734374046,
-0.03418504819273949,
0.06250602751970291,
-0.02524980530142784,
0.05647341534495354,
0.05115263909101486,
0.18922631442546844,
0.12448123097419739,
0.003569660009816289,
0.069115549325943,
0.055840928107500076,
-0.12178240716457367,
-0.11671214550733566,
0.1077936589717865,
-0.06371849775314331,
0.10950656980276108,
-0.06233043223619461,
0.18205806612968445,
0.0656684935092926,
-0.19160066545009613,
0.017418725416064262,
-0.05169399082660675,
-0.10679478943347931,
-0.0937274843454361,
-0.08642249554395676,
-0.0895860493183136,
-0.09895248711109161,
0.01973528414964676,
-0.11934094876050949,
0.037944212555885315,
0.11105069518089294,
0.02984197996556759,
0.012635610066354275,
0.10692594200372696,
-0.013212772086262703,
-0.00016057072207331657,
0.06588426232337952,
0.03940131887793541,
0.03514843434095383,
-0.06886350363492966,
-0.09671371430158615,
0.01947573758661747,
-0.01193066593259573,
0.06848926097154617,
-0.018126748502254486,
0.06201893836259842,
0.06400986015796661,
0.0029265349730849266,
-0.07534647732973099,
0.03973058611154556,
-0.0026777866296470165,
0.027966516092419624,
0.09570958465337753,
0.02829769253730774,
0.0141080217435956,
-0.019732141867280006,
0.2283022701740265,
-0.06180987134575844,
-0.08317163586616516,
-0.14410898089408875,
0.15637601912021637,
-0.017730630934238434,
-0.01868053525686264,
0.07327253371477127,
-0.09686097502708435,
-0.01044459268450737,
0.17274409532546997,
0.15730510652065277,
-0.09534463286399841,
-0.01660751737654209,
0.01294760126620531,
-0.016628902405500412,
-0.030836693942546844,
0.12377694994211197,
0.07051742821931839,
0.03343365713953972,
-0.06947138160467148,
-0.0009570361580699682,
-0.004556144122034311,
-0.03275369480252266,
-0.061556510627269745,
0.10444584488868713,
-0.009631427004933357,
-0.015451516956090927,
-0.040422871708869934,
0.05243653431534767,
-0.047527141869068146,
-0.11578447371721268,
0.06193634495139122,
-0.1618393063545227,
-0.1767985075712204,
-0.007249937392771244,
0.04074837267398834,
-0.04822336882352829,
0.04377923905849457,
-0.02128954418003559,
-0.01878274232149124,
0.12002873420715332,
-0.03369060158729553,
-0.027808576822280884,
-0.05833979323506355,
0.040225133299827576,
-0.06500659883022308,
0.20613473653793335,
-0.0012199267512187362,
0.07646220177412033,
0.11742639541625977,
0.05305199697613716,
-0.13505828380584717,
0.01645505428314209,
0.08923781663179398,
-0.07726160436868668,
0.03879941999912262,
0.12458851933479309,
-0.045847080647945404,
0.07614164054393768,
0.06504753232002258,
-0.08058003336191177,
0.0025615242775529623,
-0.04543599113821983,
-0.0034909371752291918,
-0.08351614326238632,
-0.03784510865807533,
-0.06523773819208145,
0.17302413284778595,
0.1703263521194458,
-0.04005636274814606,
0.0025901529006659985,
-0.033199507743120193,
0.025771362707018852,
0.025504201650619507,
0.05907554179430008,
-0.021731188520789146,
-0.2007506638765335,
0.024366363883018494,
0.07063966989517212,
0.05293986573815346,
-0.20490127801895142,
-0.08001872152090073,
-0.003700708271935582,
-0.05200023949146271,
-0.09618343412876129,
0.12253080308437347,
0.011269175447523594,
0.013005057349801064,
-0.03406668081879616,
-0.07181177288293839,
-0.03603520616889,
0.12672223150730133,
-0.1336887776851654,
-0.07445097714662552
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-large-cased-lora-1.58M-snli-model3
This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8051
- Accuracy: 0.6975
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 256
- eval_batch_size: 256
- seed: 74
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.5037 | 1.0 | 2146 | 0.4157 | 0.8407 |
| 0.4587 | 2.0 | 4292 | 0.3823 | 0.8574 |
| 0.446 | 3.0 | 6438 | 0.3734 | 0.8612 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-large-cased", "model-index": [{"name": "bert-large-cased-lora-1.58M-snli-model3", "results": []}]} | text-classification | varun-v-rao/bert-large-cased-lora-1.58M-snli-model3 | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:bert-large-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T08:04:36+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-large-cased-lora-1.58M-snli-model3
=======================================
This model is a fine-tuned version of bert-large-cased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8051
* Accuracy: 0.6975
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 256
* eval\_batch\_size: 256
* seed: 74
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 74\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 74\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
68,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-large-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 256\n* eval\\_batch\\_size: 256\n* seed: 74\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.08567966520786285,
0.09621097147464752,
-0.0021669145207852125,
0.1123308539390564,
0.14731737971305847,
0.0235870610922575,
0.1576298177242279,
0.10887430608272552,
-0.07177284359931946,
0.04974283277988434,
0.12772446870803833,
0.125077486038208,
0.012325609102845192,
0.12008026242256165,
-0.0652211457490921,
-0.22886139154434204,
0.0040700482204556465,
0.03652415797114372,
-0.06846872717142105,
0.11444934457540512,
0.09119358658790588,
-0.13064758479595184,
0.09569049626588821,
-0.00863171648234129,
-0.1864377111196518,
0.014920580200850964,
0.020959261804819107,
-0.05371030047535896,
0.13356833159923553,
0.037785161286592484,
0.13846832513809204,
0.016992740333080292,
0.09349150955677032,
-0.19954359531402588,
0.008453969843685627,
0.058402035385370255,
-0.002388069173321128,
0.08370272070169449,
0.03683341667056084,
0.020957620814442635,
0.08279800415039062,
-0.08907836675643921,
0.06192367523908615,
0.01651361584663391,
-0.11517087370157242,
-0.20018284022808075,
-0.07161746919155121,
0.04139389842748642,
0.08493775129318237,
0.07918689399957657,
-0.009384164586663246,
0.12197940796613693,
-0.05069129914045334,
0.08775658905506134,
0.2169409841299057,
-0.32037192583084106,
-0.06475020200014114,
0.055568963289260864,
0.03269252926111221,
0.08597128093242645,
-0.10369390994310379,
-0.01830681599676609,
0.07385455071926117,
0.025153890252113342,
0.13523192703723907,
-0.029325539246201515,
-0.06995304673910141,
0.009981123730540276,
-0.14889898896217346,
-0.019813446328043938,
0.15025228261947632,
0.04924806207418442,
-0.041587602347135544,
-0.05053238198161125,
-0.06406829506158829,
-0.1511784791946411,
-0.03739126771688461,
-0.02468590810894966,
0.0552375502884388,
-0.026602916419506073,
-0.06368046998977661,
-0.011974116787314415,
-0.1082015186548233,
-0.08148366212844849,
-0.06748282164335251,
0.14091931283473969,
0.03779371455311775,
0.0015866546891629696,
-0.018612012267112732,
0.09321478754281998,
-0.03647162765264511,
-0.12110982835292816,
0.022934187203645706,
0.026640329509973526,
0.008750119246542454,
-0.057729557156562805,
-0.05475214868783951,
-0.07243983447551727,
0.023025238886475563,
0.1317339986562729,
-0.053664714097976685,
0.04990854486823082,
0.014011186547577381,
0.04862474650144577,
-0.10005985200405121,
0.17541831731796265,
-0.04189220443367958,
-0.020399244502186775,
0.010198906995356083,
0.07445637136697769,
0.04167691618204117,
-0.003236187854781747,
-0.13255809247493744,
0.022162720561027527,
0.10322465002536774,
0.012440345250070095,
-0.08297666162252426,
0.07915948331356049,
-0.04673293977975845,
-0.0009025992476381361,
0.009616674855351448,
-0.08561647683382034,
0.032472528517246246,
0.003839158220216632,
-0.050915349274873734,
-0.05830736458301544,
0.02434684894979,
0.012205597013235092,
0.006005085539072752,
0.1145201250910759,
-0.09170599281787872,
0.009097161702811718,
-0.09066475927829742,
-0.12443967908620834,
0.024435941129922867,
-0.08161994069814682,
0.022028522565960884,
-0.11224399507045746,
-0.15503759682178497,
-0.002916628262028098,
0.052450429648160934,
-0.02694687806069851,
-0.030766231939196587,
-0.04976489394903183,
-0.07107491046190262,
0.01601996086537838,
-0.017390776425600052,
0.08849076181650162,
-0.06339897960424423,
0.09125755727291107,
0.04064726084470749,
0.06978190690279007,
-0.05350092053413391,
0.03617202118039131,
-0.09521683305501938,
0.028115736320614815,
-0.17399603128433228,
0.012921066023409367,
-0.06323050707578659,
0.057697366923093796,
-0.08352857083082199,
-0.07896091043949127,
-0.0065481653437018394,
0.014849095605313778,
0.06986410915851593,
0.08256863802671432,
-0.1650199592113495,
-0.06847614794969559,
0.1795961856842041,
-0.08760231733322144,
-0.14208658039569855,
0.13161785900592804,
-0.05662157014012337,
0.048831719905138016,
0.06389398127794266,
0.1756407916545868,
0.05515342205762863,
-0.09511371701955795,
-0.006566255819052458,
0.003990645986050367,
0.06135408952832222,
-0.051639899611473083,
0.06685785204172134,
0.0027653949800878763,
-0.00871659629046917,
0.014779394492506981,
-0.04427983984351158,
0.04909432679414749,
-0.07945643365383148,
-0.0839768573641777,
-0.040533922612667084,
-0.09931714832782745,
0.04620734602212906,
0.054543592035770416,
0.06767246127128601,
-0.11245090514421463,
-0.08827507495880127,
0.07151453197002411,
0.06830998510122299,
-0.07443376630544662,
0.017814159393310547,
-0.0704391747713089,
0.07895810902118683,
-0.06480098515748978,
-0.015345730818808079,
-0.15026184916496277,
-0.048146896064281464,
0.01992042362689972,
-0.007911194115877151,
0.018909616395831108,
0.010147592052817345,
0.07133549451828003,
0.0772620290517807,
-0.07072587311267853,
-0.018672777339816093,
-0.01873120851814747,
0.015551537275314331,
-0.13153532147407532,
-0.19868306815624237,
-0.0183318592607975,
-0.02953818254172802,
0.12380300462245941,
-0.23763258755207062,
0.04914534091949463,
0.0009846658213064075,
0.07838599383831024,
0.035775281488895416,
-0.0032936169300228357,
-0.05087234824895859,
0.07456505298614502,
-0.04441530629992485,
-0.05307702347636223,
0.057496219873428345,
0.006290437653660774,
-0.0879770815372467,
-0.04195857793092728,
-0.1450689435005188,
0.19228975474834442,
0.13393151760101318,
-0.0938953086733818,
-0.08177974820137024,
-0.012523651123046875,
-0.042322058230638504,
-0.028315318748354912,
-0.05258756875991821,
-0.001688243355602026,
0.13815638422966003,
-0.019513916224241257,
0.15301260352134705,
-0.08541051298379898,
-0.03672083839774132,
0.02733921818435192,
-0.04474529251456261,
0.008195677772164345,
0.10971210896968842,
0.1223260760307312,
-0.11104285717010498,
0.15235216915607452,
0.18089210987091064,
-0.1065472960472107,
0.14063866436481476,
-0.04322395846247673,
-0.0617859847843647,
-0.021949144080281258,
0.0007877471507526934,
0.008773954585194588,
0.11269398778676987,
-0.12912781536579132,
-0.0032947224099189043,
0.006100375670939684,
0.014375896193087101,
0.01526145450770855,
-0.21977022290229797,
-0.025721831247210503,
0.039911866188049316,
-0.04924981668591499,
0.011118602938950062,
-0.024748075753450394,
-0.008076448924839497,
0.10208684206008911,
-0.0011590387439355254,
-0.08877485990524292,
0.04289042949676514,
-0.003869319334626198,
-0.08560211956501007,
0.21238595247268677,
-0.07549075782299042,
-0.13820746541023254,
-0.14057056605815887,
-0.06683245301246643,
-0.03657122701406479,
0.023509785532951355,
0.05968785285949707,
-0.08089073747396469,
-0.039288073778152466,
-0.10647467523813248,
0.01097234059125185,
0.030895225703716278,
0.03851422294974327,
0.018301811069250107,
0.004821025300770998,
0.0809832289814949,
-0.10317113995552063,
-0.011920900084078312,
-0.050992779433727264,
-0.06405830383300781,
0.025990908965468407,
0.0272502638399601,
0.11892379075288773,
0.14774714410305023,
-0.03311122953891754,
-0.002952525857836008,
-0.03507140278816223,
0.21328958868980408,
-0.05644324794411659,
-0.022083431482315063,
0.12051611393690109,
-0.030936162918806076,
0.0441446490585804,
0.13508537411689758,
0.06955090165138245,
-0.09178880602121353,
0.020042110234498978,
0.0416201576590538,
-0.030921414494514465,
-0.21783484518527985,
-0.03680634871125221,
-0.03863849118351936,
0.016194229945540428,
0.1013450101017952,
0.034354131668806076,
0.0240244772285223,
0.0662519633769989,
0.03744129464030266,
0.0891467034816742,
-0.021084297448396683,
0.07013384997844696,
0.12175054848194122,
0.04266512021422386,
0.1265818029642105,
-0.049998246133327484,
-0.06707131862640381,
0.036043815314769745,
0.0033032067585736513,
0.20702849328517914,
0.029422780498862267,
0.1422756314277649,
0.05557971075177193,
0.14542779326438904,
0.0020834659226238728,
0.07276687771081924,
-0.013435911387205124,
-0.04749342426657677,
-0.014651050791144371,
-0.05092475190758705,
-0.02207321859896183,
0.041160453110933304,
-0.08806012570858002,
0.05781986936926842,
-0.11318585276603699,
0.01793339103460312,
0.058486148715019226,
0.23288705945014954,
0.04597805440425873,
-0.32553553581237793,
-0.08528964221477509,
0.023912813514471054,
-0.030737794935703278,
-0.018960075452923775,
0.03720267862081528,
0.14255636930465698,
-0.04997030273079872,
0.024617668241262436,
-0.0706368163228035,
0.07921476662158966,
-0.048097603023052216,
0.046333689242601395,
0.06900139898061752,
0.09059617668390274,
-0.00690377177670598,
0.07086719572544098,
-0.2706577479839325,
0.2769985496997833,
0.012780082412064075,
0.06746049225330353,
-0.05511727184057236,
0.003310876665636897,
0.030799726024270058,
0.08142697066068649,
0.07461018115282059,
-0.02293369174003601,
-0.05998317152261734,
-0.20179162919521332,
-0.05283363536000252,
0.028486784547567368,
0.09846168011426926,
-0.031455084681510925,
0.09821416437625885,
-0.034813493490219116,
0.0018854702357202768,
0.09192727506160736,
-0.00785974133759737,
-0.07272341102361679,
-0.09874488413333893,
-0.020077919587492943,
0.0470941960811615,
-0.04221734404563904,
-0.07950098812580109,
-0.1054529994726181,
-0.13766071200370789,
0.15173247456550598,
-0.05867333710193634,
-0.019711168482899666,
-0.09170735627412796,
0.07004556059837341,
0.04952486976981163,
-0.07781539112329483,
0.05055055022239685,
0.008947314694523811,
0.09103373438119888,
0.021640921011567116,
-0.05789864808320999,
0.12950226664543152,
-0.07504080981016159,
-0.16975747048854828,
-0.07795125246047974,
0.09752190858125687,
0.01815592683851719,
0.047190308570861816,
-0.0020851485896855593,
0.0032426153775304556,
-0.007041108328849077,
-0.08117342740297318,
0.023322327062487602,
-0.004789724946022034,
0.0635828822851181,
0.011346287094056606,
-0.07892056554555893,
-0.006329200230538845,
-0.045340560376644135,
-0.03127651661634445,
0.16831032931804657,
0.2648019790649414,
-0.09310854226350784,
-0.006641915999352932,
0.059809911996126175,
-0.0724799782037735,
-0.21103116869926453,
0.0443609319627285,
0.03468228504061699,
0.004837602376937866,
0.04024973511695862,
-0.14794255793094635,
0.1280824840068817,
0.10760331153869629,
-0.023610349744558334,
0.10250906646251678,
-0.2880629301071167,
-0.13458938896656036,
0.13076238334178925,
0.15646910667419434,
0.11119473725557327,
-0.15471217036247253,
-0.032148461788892746,
-0.03586883842945099,
-0.10530194640159607,
0.11618297547101974,
-0.12279820442199707,
0.11284962296485901,
-0.00914222002029419,
0.05463975667953491,
0.0017892553005367517,
-0.05931461974978447,
0.12006688863039017,
-0.012043984606862068,
0.11033231019973755,
-0.05929532274603844,
-0.03068634867668152,
0.044175777584314346,
-0.049520645290613174,
0.023597149178385735,
-0.12357097864151001,
0.028019974008202553,
-0.05003464221954346,
-0.03660937771201134,
-0.04325567185878754,
0.03476659581065178,
-0.03687603771686554,
-0.06667692214250565,
-0.04201366752386093,
0.024768821895122528,
0.03460526466369629,
-0.014633278362452984,
0.13810966908931732,
0.014363455586135387,
0.1436259150505066,
0.13751043379306793,
0.08780384808778763,
-0.06373384594917297,
-0.03011765144765377,
0.0023190623614937067,
-0.03689198940992355,
0.0688224732875824,
-0.14884530007839203,
0.03838150575757027,
0.12540505826473236,
0.005379671696573496,
0.14968565106391907,
0.08079095929861069,
-0.02326544001698494,
0.00744197191670537,
0.0699269250035286,
-0.1585760861635208,
-0.10711181163787842,
-0.0010261392453685403,
-0.033326517790555954,
-0.12358292192220688,
0.06985628604888916,
0.11440018564462662,
-0.07640695571899414,
0.008951093070209026,
-0.005684601608663797,
0.008388746529817581,
-0.04161989688873291,
0.1784776896238327,
0.06596875190734863,
0.04779478535056114,
-0.072739377617836,
0.07613913714885712,
0.044300977140665054,
-0.07450323551893234,
0.01767042838037014,
0.04148894175887108,
-0.0803937241435051,
-0.03964000195264816,
0.05628412589430809,
0.19876891374588013,
-0.02804100140929222,
-0.05935779586434364,
-0.14236879348754883,
-0.12226063013076782,
0.058013785630464554,
0.19258694350719452,
0.1019933819770813,
0.009079438634216785,
-0.041318513453006744,
0.021815462037920952,
-0.11936652660369873,
0.1166028082370758,
0.02517736330628395,
0.08811108767986298,
-0.1546163707971573,
0.1244887039065361,
-0.0023686469066888094,
0.003924370277673006,
-0.025227440521121025,
0.05379749462008476,
-0.12633268535137177,
-0.008163939230144024,
-0.13094709813594818,
-0.006461053155362606,
-0.020541973412036896,
0.005161628592759371,
0.007910077460110188,
-0.04893519729375839,
-0.06407655030488968,
0.011043757200241089,
-0.09938110411167145,
-0.016321511939167976,
0.04198840633034706,
0.06793850660324097,
-0.12220326066017151,
-0.037308771163225174,
0.023868907243013382,
-0.06347210705280304,
0.06648027151823044,
0.011347840540111065,
0.025678927078843117,
0.05339162051677704,
-0.17471416294574738,
0.03583073243498802,
0.07391393929719925,
0.01616944931447506,
0.047253284603357315,
-0.08501579612493515,
-0.009821335785090923,
-0.0031202968675643206,
0.05023357272148132,
0.020565157756209373,
0.08063828200101852,
-0.12672050297260284,
0.010712743736803532,
-0.03166966512799263,
-0.06654544174671173,
-0.050144873559474945,
0.021870676428079605,
0.08513451367616653,
-0.0016492238501086831,
0.1990468055009842,
-0.0972619280219078,
0.012177364900708199,
-0.2040526419878006,
0.011271581053733826,
0.0038448930718004704,
-0.11145354807376862,
-0.12500335276126862,
-0.06007762625813484,
0.04798656702041626,
-0.06372922658920288,
0.15573875606060028,
0.020668327808380127,
0.010394959710538387,
0.03640100359916687,
-0.04051535576581955,
0.03941371664404869,
0.031639739871025085,
0.2293390929698944,
0.02994988113641739,
-0.046060413122177124,
0.016295790672302246,
0.03289628028869629,
0.11738844960927963,
0.08963257074356079,
0.17746827006340027,
0.16360999643802643,
-0.05304712429642677,
0.10682712495326996,
0.05405328422784805,
-0.054004184901714325,
-0.14257656037807465,
0.06293654441833496,
-0.04844549298286438,
0.10251419246196747,
-0.02595582976937294,
0.21493886411190033,
0.08658701926469803,
-0.1542416661977768,
0.008111810311675072,
-0.058300819247961044,
-0.08382443338632584,
-0.11341633647680283,
-0.049431875348091125,
-0.10055463016033173,
-0.14513233304023743,
0.002016722923144698,
-0.11553353071212769,
0.00033263119985349476,
0.09961424767971039,
0.006182770244777203,
-0.015990307554602623,
0.17697204649448395,
0.0015996228903532028,
0.04022565856575966,
0.05901065468788147,
0.006125730462372303,
-0.031403798609972,
-0.10162457823753357,
-0.0865362137556076,
-0.00558810168877244,
-0.021162796765565872,
0.01996808685362339,
-0.05184285342693329,
-0.025480400770902634,
0.038466617465019226,
-0.003835709998384118,
-0.0952024981379509,
0.006923345848917961,
0.01974424719810486,
0.05293894186615944,
0.03428436443209648,
0.0029303240589797497,
0.0059653231874108315,
-0.00046705995919182897,
0.19806544482707977,
-0.07934663444757462,
-0.05843416228890419,
-0.0955081656575203,
0.21335528790950775,
0.023373395204544067,
0.02458437718451023,
0.011278162710368633,
-0.08452268689870834,
0.02042376808822155,
0.23236091434955597,
0.1873224824666977,
-0.08065836131572723,
0.0005815267213620245,
0.0009786031441763043,
-0.010526109486818314,
-0.037836041301488876,
0.10235595703125,
0.11617212742567062,
0.03237270191311836,
-0.07524263113737106,
-0.051808975636959076,
-0.03255174309015274,
0.0008920197724364698,
-0.03489512950181961,
0.05398619920015335,
0.04101349040865898,
0.012704310938715935,
-0.051886122673749924,
0.04631844535470009,
-0.03139922395348549,
-0.10607399046421051,
0.06299348175525665,
-0.1977020800113678,
-0.1490129977464676,
-0.00434118090197444,
0.12445677816867828,
-0.020951831713318825,
0.05204864591360092,
-0.03392133116722107,
-0.006527419202029705,
0.06386896967887878,
-0.025099031627178192,
-0.07764465361833572,
-0.07043294608592987,
0.058435458689928055,
-0.08776001632213593,
0.23614253103733063,
-0.03755934536457062,
0.05970388278365135,
0.13139846920967102,
0.0436905100941658,
-0.0674576386809349,
0.07952137291431427,
0.04661836475133896,
-0.07642731070518494,
0.026229318231344223,
0.0675322487950325,
-0.044531360268592834,
0.12252666056156158,
0.053744252771139145,
-0.13648569583892822,
0.017033278942108154,
-0.057785797864198685,
-0.08745232224464417,
-0.059392932802438736,
-0.037183474749326706,
-0.06564335525035858,
0.1348012387752533,
0.19449321925640106,
-0.02851918525993824,
0.002466397825628519,
-0.04397742450237274,
0.02702910080552101,
0.06862524896860123,
0.0381658636033535,
-0.03293226286768913,
-0.23239803314208984,
0.026530539616942406,
0.07142167538404465,
-0.008098815567791462,
-0.2833006978034973,
-0.09095381200313568,
-0.00021167616068851203,
-0.045675329864025116,
-0.1003611609339714,
0.07491831481456757,
0.1305171400308609,
0.05826713144779205,
-0.06619168072938919,
-0.10228752344846725,
-0.07443390041589737,
0.15132254362106323,
-0.13403356075286865,
-0.10304783284664154
] |
null | null | transformers | # merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the SLERP merge method.
### Models Merged
The following models were included in the merge:
* [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B)
* [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
merge_method: slerp
base_model: mistralai/Mistral-7B-Instruct-v0.2
slices:
- sources:
- model: mistralai/Mistral-7B-Instruct-v0.2
layer_range: [0, 32]
- model: teknium/OpenHermes-2.5-Mistral-7B
layer_range: [0, 32]
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5 # fallback for rest of tensors
dtype: bfloat16
```
| {"library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["teknium/OpenHermes-2.5-Mistral-7B", "mistralai/Mistral-7B-Instruct-v0.2"]} | text-generation | chanwit/flux-7b-base-stage-00 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:teknium/OpenHermes-2.5-Mistral-7B",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T08:04:40+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #mergekit #merge #conversational #base_model-teknium/OpenHermes-2.5-Mistral-7B #base_model-mistralai/Mistral-7B-Instruct-v0.2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # merge
This is a merge of pre-trained language models created using mergekit.
## Merge Details
### Merge Method
This model was merged using the SLERP merge method.
### Models Merged
The following models were included in the merge:
* teknium/OpenHermes-2.5-Mistral-7B
* mistralai/Mistral-7B-Instruct-v0.2
### Configuration
The following YAML configuration was used to produce this model:
| [
"# merge\n\nThis is a merge of pre-trained language models created using mergekit.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the SLERP merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* teknium/OpenHermes-2.5-Mistral-7B\n* mistralai/Mistral-7B-Instruct-v0.2",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #conversational #base_model-teknium/OpenHermes-2.5-Mistral-7B #base_model-mistralai/Mistral-7B-Instruct-v0.2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# merge\n\nThis is a merge of pre-trained language models created using mergekit.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the SLERP merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* teknium/OpenHermes-2.5-Mistral-7B\n* mistralai/Mistral-7B-Instruct-v0.2",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
95,
18,
4,
18,
44,
17
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #conversational #base_model-teknium/OpenHermes-2.5-Mistral-7B #base_model-mistralai/Mistral-7B-Instruct-v0.2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# merge\n\nThis is a merge of pre-trained language models created using mergekit.## Merge Details### Merge Method\n\nThis model was merged using the SLERP merge method.### Models Merged\n\nThe following models were included in the merge:\n* teknium/OpenHermes-2.5-Mistral-7B\n* mistralai/Mistral-7B-Instruct-v0.2### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.09678915143013,
-0.09477753937244415,
0.00009844549640547484,
0.00583726167678833,
0.11328937113285065,
0.07891891151666641,
0.19403286278247833,
0.033664681017398834,
0.17628948390483856,
0.0029148501344025135,
-0.004471074789762497,
0.05555838719010353,
0.02571053057909012,
0.1751854121685028,
-0.05558052286505699,
-0.24962931871414185,
0.0876036137342453,
-0.05943034216761589,
-0.1415245532989502,
0.08439987897872925,
0.11670073121786118,
-0.03804910555481911,
0.09960071742534637,
0.023733410984277725,
-0.13283033668994904,
0.0015747644938528538,
-0.005880521610379219,
0.0077203260734677315,
0.1395859271287918,
0.11367758363485336,
0.07023179531097412,
0.015334757044911385,
0.021432118490338326,
-0.12155956774950027,
0.051998794078826904,
-0.006494148168712854,
0.006728685460984707,
0.059279028326272964,
0.049740020185709,
-0.009009228087961674,
0.11354166269302368,
-0.0454472154378891,
0.037570953369140625,
0.09373632818460464,
-0.09405987709760666,
-0.11321613937616348,
-0.10823865234851837,
0.06434396654367447,
0.12474746257066727,
0.021125035360455513,
-0.015572271309792995,
0.10801874846220016,
0.056475479155778885,
0.07985066622495651,
0.053051553666591644,
-0.2525527775287628,
-0.015148038975894451,
0.18290196359157562,
0.08370159566402435,
-0.0336129255592823,
0.026438023895025253,
0.052141301333904266,
0.06283587217330933,
-0.019374433904886246,
0.0031311367638409138,
-0.056715548038482666,
0.1419130116701126,
-0.08972921967506409,
-0.15674038231372833,
0.020012548193335533,
0.14601030945777893,
0.01590667851269245,
-0.031942784786224365,
-0.12447338551282883,
-0.15006467700004578,
0.0990859791636467,
-0.02137189731001854,
-0.08417104184627533,
0.01375366747379303,
0.025131413713097572,
0.1409582793712616,
-0.08613754063844681,
-0.045501627027988434,
-0.019507059827446938,
-0.11380728334188461,
0.1829819679260254,
0.024832800030708313,
0.02588300034403801,
-0.08439049869775772,
0.05669640377163887,
-0.22234100103378296,
-0.09140941500663757,
-0.007519913371652365,
-0.09355614334344864,
-0.08959893137216568,
-0.04002273082733154,
-0.11896222084760666,
-0.1696554571390152,
0.06473653763532639,
0.21259301900863647,
-0.10521840304136276,
0.02611387148499489,
0.07926502823829651,
0.06174676492810249,
0.04761619120836258,
-0.013525577262043953,
-0.15237605571746826,
-0.13482527434825897,
0.032377708703279495,
-0.007890083827078342,
0.10300298780202866,
-0.00630947295576334,
-0.11837731301784515,
-0.04187542572617531,
-0.04223838075995445,
-0.02017902210354805,
0.026924559846520424,
0.11952091753482819,
-0.03162071108818054,
-0.07878708839416504,
0.17012085020542145,
-0.06474146991968155,
-0.00009050927474163473,
-0.010314330458641052,
-0.02353673055768013,
0.016384344547986984,
0.12813933193683624,
0.06993532925844193,
0.05801166966557503,
0.04254385456442833,
-0.06335272639989853,
-0.008682815358042717,
-0.05151347815990448,
-0.12410596758127213,
0.0029698561411350965,
-0.03147037327289581,
-0.02846125327050686,
-0.11945050209760666,
-0.22119548916816711,
-0.010533473454415798,
0.05038017779588699,
-0.03283808007836342,
0.027163727208971977,
-0.01840836927294731,
0.04041594639420509,
0.000010987122550432105,
0.014597936533391476,
-0.0462183952331543,
-0.004852385725826025,
-0.05752629414200783,
-0.0439908467233181,
0.05852268636226654,
-0.19671545922756195,
0.020213371142745018,
-0.08707372099161148,
0.11688496172428131,
-0.16512048244476318,
0.09680737555027008,
-0.006510311737656593,
0.06100713834166527,
-0.12417105585336685,
-0.006194200366735458,
0.01201990433037281,
0.024792753159999847,
0.08530090004205704,
0.17016108334064484,
-0.1455734670162201,
-0.05915499106049538,
0.11016056686639786,
-0.15164358913898468,
-0.15025432407855988,
0.12233826518058777,
-0.006715389899909496,
0.12582182884216309,
0.03537692874670029,
0.18479040265083313,
0.13629528880119324,
-0.025443123653531075,
0.03053201362490654,
-0.02218085341155529,
-0.001340943737886846,
0.05193747580051422,
0.07494078576564789,
-0.02338327094912529,
-0.09309667348861694,
0.03401850536465645,
0.001014228444546461,
0.12678144872188568,
-0.04376678168773651,
-0.05587374046444893,
-0.0695272758603096,
-0.061938945204019547,
0.17439106106758118,
-0.031025879085063934,
0.03874114900827408,
-0.059437867254018784,
-0.0031108935363590717,
0.13345526158809662,
0.11935771256685257,
-0.056057486683130264,
0.017718041315674782,
-0.03333624452352524,
0.1264340877532959,
-0.07728467881679535,
0.030055273324251175,
-0.13034680485725403,
-0.12222298979759216,
-0.02955072559416294,
-0.05825120210647583,
0.012689466588199139,
0.056288979947566986,
0.07748295366764069,
0.07238046079874039,
-0.06323038786649704,
-0.007295937743037939,
0.10793773829936981,
0.04620939865708351,
-0.03909667953848839,
-0.1893734633922577,
-0.10370100289583206,
-0.09153692424297333,
0.2965151071548462,
-0.10733065754175186,
0.08789710700511932,
-0.0748700499534607,
0.19794200360774994,
-0.03363548964262009,
0.031004998832941055,
0.045227136462926865,
0.00939419399946928,
-0.04438398778438568,
-0.0045411912724375725,
0.08736424148082733,
0.010489784181118011,
-0.20694875717163086,
0.12807859480381012,
-0.1754320114850998,
-0.01953805983066559,
0.07234973460435867,
0.01672808825969696,
-0.053324513137340546,
-0.12405319511890411,
-0.020535411313176155,
-0.07579947263002396,
0.06425724178552628,
-0.08664172887802124,
0.08937612920999527,
0.015302607789635658,
0.10037054121494293,
-0.017644129693508148,
-0.010922551155090332,
0.0241885706782341,
-0.04969530925154686,
-0.07679330557584763,
0.03303117677569389,
-0.029482915997505188,
-0.27086812257766724,
0.08982526510953903,
0.2041412740945816,
0.028088418766856194,
0.088246650993824,
0.042663224041461945,
0.021861344575881958,
-0.06535275280475616,
-0.018371880054473877,
-0.01971902698278427,
-0.02372671663761139,
-0.07837342470884323,
0.04662368446588516,
0.06044028326869011,
0.0002708498213905841,
0.038092516362667084,
-0.0659523606300354,
0.036976784467697144,
0.05363311618566513,
0.034721605479717255,
0.0821424126625061,
0.1445714384317398,
-0.01342503447085619,
0.060095611959695816,
0.016159523278474808,
-0.039544083178043365,
0.018538180738687515,
-0.014312143437564373,
-0.14716577529907227,
0.17200252413749695,
-0.1352710872888565,
-0.13068309426307678,
-0.14948399364948273,
-0.09255097061395645,
-0.09773074090480804,
-0.0033864853903651237,
0.045397233217954636,
0.005243992432951927,
-0.0525173656642437,
-0.08589547872543335,
0.10659381747245789,
0.051250945776700974,
-0.009104276075959206,
0.022060327231884003,
-0.05544891208410263,
0.00648093456402421,
-0.10421839356422424,
-0.01729230210185051,
-0.007645782083272934,
-0.029651863500475883,
0.05073487013578415,
-0.05211710184812546,
0.084537073969841,
0.1738233119249344,
-0.01965481974184513,
-0.03378639370203018,
-0.030692053958773613,
0.19966870546340942,
-0.03729651868343353,
0.11868566274642944,
0.1685430407524109,
-0.10151185095310211,
0.05861826613545418,
0.2563093304634094,
0.03493213281035423,
-0.04022757336497307,
0.01803460717201233,
-0.06819261610507965,
-0.06818599253892899,
-0.15894801914691925,
-0.16186755895614624,
-0.09179198741912842,
0.009874841198325157,
-0.0036834825295954943,
0.007541699334979057,
0.09649128466844559,
0.07325573265552521,
-0.08170424401760101,
-0.037880536168813705,
0.07794340699911118,
0.06710927188396454,
0.18531857430934906,
0.0004857271269429475,
0.09249722212553024,
-0.031470995396375656,
0.033777520060539246,
0.07017392665147781,
-0.023598961532115936,
0.13376270234584808,
0.06572887301445007,
0.13786625862121582,
0.08716034144163132,
-0.009213779121637344,
0.04969288781285286,
0.07369955629110336,
-0.0393185019493103,
0.016192683950066566,
-0.031153442338109016,
-0.10242011398077011,
0.017142631113529205,
0.10022150725126266,
-0.06806305050849915,
0.09833154827356339,
-0.07732220739126205,
0.007768075913190842,
0.06069353222846985,
0.12283489108085632,
0.1185128465294838,
-0.2770572304725647,
-0.17149916291236877,
0.08091530948877335,
0.06043623015284538,
0.004008646123111248,
-0.047429583966732025,
0.05895298719406128,
-0.08329465985298157,
0.2010468691587448,
-0.06919647753238678,
0.08869510143995285,
0.0791805163025856,
0.0014440844533964992,
0.008948598988354206,
0.13385836780071259,
0.0015216373139992356,
0.018630634993314743,
-0.1323644369840622,
0.18878352642059326,
0.021313007920980453,
-0.05914741009473801,
0.014112966135144234,
0.03421185910701752,
0.06188680976629257,
0.261589914560318,
0.018676847219467163,
0.014683757908642292,
0.009355111047625542,
-0.022362016141414642,
-0.07060888409614563,
-0.023202314972877502,
-0.00499004777520895,
-0.04425790533423424,
0.07994838058948517,
-0.06021199747920036,
-0.05113977938890457,
0.035841021686792374,
0.18108701705932617,
-0.141445592045784,
-0.1489051729440689,
0.0793314203619957,
0.06879544258117676,
0.040009599179029465,
-0.061135802417993546,
-0.06733449548482895,
-0.13509626686573029,
0.23188036680221558,
0.008511505089700222,
-0.09046492725610733,
-0.10933253914117813,
0.005971051286906004,
0.13209064304828644,
-0.06762444972991943,
0.08126436173915863,
-0.02340508997440338,
0.08120261132717133,
-0.09726178646087646,
-0.17343884706497192,
0.13256990909576416,
-0.12024865299463272,
-0.08771219104528427,
-0.021921297535300255,
0.08831967413425446,
-0.011265073902904987,
0.023836582899093628,
0.00008873250044416636,
0.05630672350525856,
-0.018656950443983078,
-0.038603171706199646,
-0.03990502655506134,
0.3174465596675873,
-0.009962753392755985,
0.10562503337860107,
-0.05550376698374748,
-0.20639309287071228,
0.0008531035273335874,
-0.012106776237487793,
0.1528834104537964,
0.2120688259601593,
-0.07180839031934738,
0.08088704198598862,
0.161336749792099,
-0.07267457991838455,
-0.17031987011432648,
0.02482728287577629,
-0.02691752091050148,
0.032840874046087265,
0.04519656300544739,
0.009044814854860306,
0.007825932465493679,
0.0725463479757309,
-0.0033768254797905684,
0.04820261895656586,
-0.33694007992744446,
-0.18245242536067963,
0.04568575695157051,
0.04753224179148674,
0.2220754474401474,
-0.07826552540063858,
-0.07232940942049026,
-0.07521944493055344,
-0.209107905626297,
-0.018693864345550537,
-0.1461620181798935,
0.0466158464550972,
-0.0005225951899774373,
0.01490371860563755,
0.016054362058639526,
-0.041838329285383224,
0.1547386795282364,
-0.025274500250816345,
0.02556304819881916,
-0.10674013197422028,
-0.014788733795285225,
0.09452259540557861,
-0.055316805839538574,
0.08900632709264755,
-0.008618495427072048,
0.012698336504399776,
0.012367600575089455,
-0.031231986358761787,
-0.04006632789969444,
0.06241655349731445,
-0.031224019825458527,
-0.08069898933172226,
-0.036136649549007416,
0.05015097185969353,
0.017755748704075813,
-0.012901396490633488,
0.10939069837331772,
-0.03710472211241722,
0.16088713705539703,
0.1843414306640625,
0.1042652502655983,
-0.04482719674706459,
0.016407327726483345,
0.024616194888949394,
-0.05846307799220085,
0.06478241831064224,
-0.07507850974798203,
-0.02729940414428711,
0.09860394895076752,
-0.008831650018692017,
0.1075073778629303,
0.03546622022986412,
-0.03042234480381012,
0.03706151992082596,
0.0501239188015461,
-0.1332402527332306,
-0.33690667152404785,
-0.019097071141004562,
0.035182297229766846,
-0.03768453374505043,
0.09853273630142212,
0.21198837459087372,
-0.09666985273361206,
-0.003923878539353609,
0.01916845701634884,
0.0277763232588768,
-0.08312695473432541,
0.08788759261369705,
-0.03895223140716553,
0.018288766965270042,
-0.07700784504413605,
0.05441495031118393,
0.053291719406843185,
-0.10832560062408447,
-0.01741841249167919,
0.07810556888580322,
-0.1453522890806198,
-0.09873753041028976,
-0.1228778064250946,
0.12964987754821777,
-0.09353157877922058,
-0.04647935554385185,
-0.06446507573127747,
-0.10564801841974258,
-0.0159294530749321,
0.1237712949514389,
0.08384832739830017,
-0.005903780460357666,
-0.015211290679872036,
-0.05359337106347084,
-0.03125888109207153,
0.05963965132832527,
0.03447653353214264,
0.09477667510509491,
-0.11412623524665833,
0.07235707342624664,
-0.03473289683461189,
0.045186884701251984,
-0.0594199113547802,
-0.0530293732881546,
-0.08412512391805649,
-0.04074173420667648,
-0.20680078864097595,
-0.024883713573217392,
-0.15895812213420868,
-0.030849315226078033,
-0.02655193954706192,
-0.05054168403148651,
0.000545418297406286,
0.026070619001984596,
-0.023431919515132904,
-0.024290116503834724,
-0.03199227154254913,
0.04327606037259102,
-0.07635848969221115,
-0.03778880462050438,
0.0014068378368392587,
-0.03992295637726784,
0.09596682339906693,
0.027353713288903236,
-0.0591190867125988,
-0.03794199600815773,
-0.15132060647010803,
-0.040395528078079224,
0.05441204831004143,
-0.002394499257206917,
0.044788360595703125,
-0.1406092792749405,
-0.0553443618118763,
-0.020866837352514267,
-0.024168267846107483,
-0.025859957560896873,
0.07197295129299164,
-0.03388645127415657,
0.04556047543883324,
0.0016214112984016538,
-0.002103495877236128,
-0.04798019677400589,
-0.04269464686512947,
0.015012001618742943,
0.07077307999134064,
0.10614252835512161,
-0.05676612630486488,
0.06514903157949448,
-0.16260528564453125,
-0.03374268487095833,
0.0030423467978835106,
-0.08741920441389084,
-0.04574592784047127,
-0.08606378734111786,
0.001468271017074585,
-0.026798320934176445,
0.12091315537691116,
-0.056575383991003036,
-0.023484015837311745,
0.03969845548272133,
-0.008707454428076744,
0.036099448800086975,
0.05559888482093811,
0.18380899727344513,
0.033501263707876205,
0.031078074127435684,
-0.09362233430147171,
0.08587931841611862,
0.018011264503002167,
0.006423084530979395,
0.09279105812311172,
0.06050762161612511,
-0.014558440074324608,
0.08052236586809158,
0.04351586475968361,
-0.0006248748977668583,
-0.029048269614577293,
-0.13848891854286194,
-0.005610278341919184,
0.044528599828481674,
-0.04398149996995926,
0.10927936434745789,
0.14149874448776245,
-0.16666896641254425,
0.07448991388082504,
0.0010278270347043872,
-0.034173883497714996,
-0.0930989682674408,
-0.08311320841312408,
-0.0743088647723198,
-0.13637255132198334,
-0.07019806653261185,
-0.07585272938013077,
-0.027041226625442505,
-0.02634831890463829,
0.006802563555538654,
0.001131126075051725,
0.15427075326442719,
-0.05486651137471199,
-0.024674976244568825,
0.004710484761744738,
-0.011459591798484325,
0.014405693858861923,
-0.011730455793440342,
-0.06651914864778519,
0.04758240655064583,
-0.02286127768456936,
-0.03142021968960762,
0.0458991639316082,
0.04289829358458519,
0.03377259895205498,
-0.024977952241897583,
-0.09755729883909225,
-0.012165989726781845,
0.07757128030061722,
0.008011406287550926,
-0.07761973887681961,
0.06767306476831436,
-0.03274572640657425,
-0.011378871276974678,
0.0871058776974678,
-0.04845418408513069,
-0.08593831956386566,
-0.08789302408695221,
0.20419487357139587,
-0.052230916917324066,
0.03865145891904831,
0.03427385911345482,
-0.09881032258272171,
-0.004370814189314842,
0.07757718861103058,
0.35807669162750244,
-0.015884578227996826,
0.010508609935641289,
-0.0019234399078413844,
0.012424771673977375,
0.01429341733455658,
0.07951332628726959,
-0.005884579848498106,
0.13026851415634155,
-0.04872399941086769,
0.11572804301977158,
-0.0030254055745899677,
-0.07687653601169586,
-0.031869251281023026,
0.013046140782535076,
-0.027218660339713097,
-0.03799564763903618,
0.03842901438474655,
0.10008405148983002,
-0.04186122119426727,
-0.045378584414720535,
-0.006978034973144531,
-0.15221919119358063,
-0.0824507474899292,
-0.0792442262172699,
0.06797929853200912,
0.029079053550958633,
0.07902843505144119,
-0.04520907625555992,
-0.010524489916861057,
0.20663930475711823,
-0.00371623900718987,
-0.08110947161912918,
-0.11879553645849228,
0.056792113929986954,
-0.021966515108942986,
-0.005331389605998993,
-0.021231040358543396,
0.048380572348833084,
0.0969310998916626,
0.013552386313676834,
-0.1073160395026207,
-0.029922684654593468,
0.05612191930413246,
0.02132612094283104,
0.03207853436470032,
0.027480140328407288,
-0.052412524819374084,
0.05658036842942238,
0.03202250599861145,
-0.24227429926395416,
0.05208210647106171,
0.005128674209117889,
-0.07397835701704025,
-0.058005403727293015,
0.08984744548797607,
-0.03314341977238655,
0.1581258773803711,
0.14763163030147552,
-0.0614541694521904,
-0.0008269738173112273,
0.00029613557853735983,
0.037886880338191986,
0.08670207858085632,
0.1157623827457428,
-0.043863218277692795,
-0.21521548926830292,
0.016387050971388817,
0.004255904816091061,
0.06965914368629456,
-0.2535237967967987,
-0.12512753903865814,
-0.10239394754171371,
-0.019501589238643646,
-0.026500951498746872,
0.10877733677625656,
0.16717635095119476,
0.010004960000514984,
-0.011857463978230953,
-0.1718357801437378,
-0.009470042772591114,
0.10540397465229034,
-0.09050586819648743,
-0.08652475476264954
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# aragpt2-base-saadeh-full2
This model is a fine-tuned version of [aubmindlab/aragpt2-base](https://huggingface.co/aubmindlab/aragpt2-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 3.9358
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 266 | 4.2973 |
| 5.5626 | 2.0 | 532 | 4.0108 |
| 5.5626 | 3.0 | 798 | 3.9358 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "base_model": "aubmindlab/aragpt2-base", "model-index": [{"name": "aragpt2-base-saadeh-full2", "results": []}]} | text-generation | ammarzaarour/aragpt2-base-saadeh-full2 | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:aubmindlab/aragpt2-base",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T08:05:26+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-aubmindlab/aragpt2-base #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| aragpt2-base-saadeh-full2
=========================
This model is a fine-tuned version of aubmindlab/aragpt2-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 3.9358
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-aubmindlab/aragpt2-base #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
74,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #gpt2 #text-generation #generated_from_trainer #base_model-aubmindlab/aragpt2-base #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09904191642999649,
0.07089155912399292,
-0.0025037259329110384,
0.09765670448541641,
0.13949760794639587,
0.012866944074630737,
0.16406048834323883,
0.11347787827253342,
-0.06742548197507858,
0.06454664468765259,
0.14321795105934143,
0.10407330095767975,
0.027088608592748642,
0.14990778267383575,
-0.07258570939302444,
-0.22652459144592285,
0.013288456946611404,
0.030081771314144135,
-0.04652031138539314,
0.11907365173101425,
0.08572535961866379,
-0.12639755010604858,
0.09129008650779724,
-0.021402010694146156,
-0.16178514063358307,
0.00547827547416091,
0.014426114037632942,
-0.05036018788814545,
0.1292852908372879,
0.03950856253504753,
0.12506744265556335,
0.026054158806800842,
0.06253495812416077,
-0.19373564422130585,
0.015900038182735443,
0.05727929621934891,
0.0025687855668365955,
0.0849907323718071,
0.043530046939849854,
-0.010170831345021725,
0.0999877005815506,
-0.0900251567363739,
0.05447402596473694,
0.016969285905361176,
-0.14899374544620514,
-0.18991464376449585,
-0.0639614462852478,
0.027294356375932693,
0.09864714741706848,
0.09024430811405182,
-0.03146935999393463,
0.13202105462551117,
-0.023150604218244553,
0.1032361388206482,
0.22821876406669617,
-0.30405479669570923,
-0.07409528642892838,
0.04808846116065979,
0.04133417829871178,
0.11246921122074127,
-0.10494164377450943,
-0.0003442632150836289,
0.060677189379930496,
0.031200474128127098,
0.11922366172075272,
-0.029975714161992073,
-0.014511392451822758,
0.0021640495397150517,
-0.1399955153465271,
-0.026019522920250893,
0.16273225843906403,
0.046443428844213486,
-0.04201524704694748,
-0.06751836091279984,
-0.07542897760868073,
-0.1417393982410431,
-0.040058329701423645,
-0.01734386943280697,
0.026340611279010773,
-0.0246293842792511,
-0.07596597075462341,
-0.04130316525697708,
-0.11032771319150925,
-0.0723101794719696,
-0.03117789514362812,
0.1430392861366272,
0.02406512200832367,
-0.006986699998378754,
-0.02372044324874878,
0.10681360960006714,
-0.03999914228916168,
-0.1471252590417862,
0.015958046540617943,
0.020045362412929535,
0.009699841029942036,
-0.045651521533727646,
-0.0540274940431118,
-0.1142973005771637,
0.012666239403188229,
0.1254807710647583,
-0.05658264458179474,
0.06361829489469528,
-0.006064045708626509,
0.04301724582910538,
-0.09385465085506439,
0.164048969745636,
-0.03753557428717613,
-0.04437709227204323,
0.012483937665820122,
0.10183882713317871,
0.04544736444950104,
-0.019785171374678612,
-0.12352138757705688,
0.02267003431916237,
0.12473093718290329,
0.02067353017628193,
-0.05871867761015892,
0.07056040316820145,
-0.050543323159217834,
-0.007976442575454712,
0.015544964000582695,
-0.08460493385791779,
0.018907567486166954,
-0.00434752507135272,
-0.05021543800830841,
-0.052076105028390884,
0.027494577690958977,
0.026109760627150536,
0.0010714646195992827,
0.09452055394649506,
-0.09420190751552582,
0.006400336045771837,
-0.07824376225471497,
-0.12264754623174667,
0.003512707306072116,
-0.06703180819749832,
0.019804831594228745,
-0.1114031970500946,
-0.17585739493370056,
-0.02251470275223255,
0.05020403116941452,
-0.04021722823381424,
-0.021904313936829567,
-0.07511880248785019,
-0.07736720144748688,
0.019806738942861557,
-0.009064245969057083,
0.08597584813833237,
-0.05872289091348648,
0.10348358005285263,
0.059633322060108185,
0.07875288277864456,
-0.04558951035141945,
0.030491862446069717,
-0.0949525460600853,
0.044534411281347275,
-0.19446387887001038,
0.03567150607705116,
-0.03788285702466965,
0.07011663168668747,
-0.08368762582540512,
-0.07376138120889664,
-0.024010593071579933,
0.005538588389754295,
0.08431091159582138,
0.11448180675506592,
-0.15153220295906067,
-0.07050635665655136,
0.19034557044506073,
-0.0854068323969841,
-0.1517861932516098,
0.1392420381307602,
-0.05384017899632454,
0.0592782124876976,
0.06799721717834473,
0.19871622323989868,
0.044964537024497986,
-0.08768124878406525,
0.007492807228118181,
-0.023071568459272385,
0.07444097846746445,
-0.01561474148184061,
0.07513957470655441,
0.005328930448740721,
0.0047097220085561275,
0.012062670662999153,
-0.02172190696001053,
0.050815075635910034,
-0.08422055840492249,
-0.07820252329111099,
-0.03367439657449722,
-0.10815794765949249,
0.04202321916818619,
0.04948374629020691,
0.07105115056037903,
-0.12210842221975327,
-0.09077726304531097,
0.032657310366630554,
0.06568684428930283,
-0.07982014119625092,
0.020518964156508446,
-0.06351681053638458,
0.08115500956773758,
-0.09591852128505707,
-0.011522477492690086,
-0.1484382301568985,
-0.059965457767248154,
0.007569728419184685,
0.037646546959877014,
0.010952356271445751,
0.009269239380955696,
0.09133000671863556,
0.08724891394376755,
-0.07225946336984634,
-0.041833437979221344,
-0.006586448755115271,
0.002670341869816184,
-0.13846848905086517,
-0.18355153501033783,
-0.0027735698968172073,
-0.03712555393576622,
0.16214051842689514,
-0.2529130280017853,
0.050141334533691406,
-0.003403886454179883,
0.08825409412384033,
0.05083503574132919,
-0.02932245470583439,
-0.021001357585191727,
0.05292554944753647,
-0.051229480654001236,
-0.0635986402630806,
0.05833632871508598,
0.004387593828141689,
-0.10671200603246689,
-0.030537061393260956,
-0.19165974855422974,
0.18224912881851196,
0.1279952973127365,
-0.07499883323907852,
-0.08588811755180359,
-0.011035678908228874,
-0.042865410447120667,
-0.025857489556074142,
-0.03338661789894104,
-0.0202870462089777,
0.14370033144950867,
0.0031938946340233088,
0.16069287061691284,
-0.08421117067337036,
-0.037248607724905014,
0.028552699834108353,
-0.05515376478433609,
0.0045961858704686165,
0.12291348725557327,
0.05970736965537071,
-0.10866699367761612,
0.14495104551315308,
0.16586029529571533,
-0.06698909401893616,
0.15910956263542175,
-0.019081005826592445,
-0.06679631024599075,
-0.02320820279419422,
0.02643115632236004,
0.023132339119911194,
0.10690461099147797,
-0.12557172775268555,
-0.005967456381767988,
0.008037460036575794,
0.021141214296221733,
0.02213515341281891,
-0.21050404012203217,
-0.03396126255393028,
0.047940827906131744,
-0.03995179757475853,
0.01862424612045288,
-0.014815699309110641,
-0.0105678029358387,
0.10487382113933563,
0.010905555449426174,
-0.05810530483722687,
0.028105318546295166,
-0.006461568176746368,
-0.08640741556882858,
0.20800553262233734,
-0.06102295592427254,
-0.16950371861457825,
-0.13598820567131042,
-0.05949333310127258,
-0.05369292199611664,
0.04395007714629173,
0.05966528132557869,
-0.08348521590232849,
-0.03363538905978203,
-0.11864811182022095,
0.029460759833455086,
0.018184030428528786,
0.023565124720335007,
0.01709510013461113,
-0.010698139667510986,
0.07071644812822342,
-0.08937328308820724,
-0.011302840895950794,
-0.03675713762640953,
-0.04901665076613426,
0.04488034546375275,
0.007598293945193291,
0.1188662126660347,
0.14289434254169464,
-0.011172846890985966,
0.007179404143244028,
-0.02902160957455635,
0.24914853274822235,
-0.08281257748603821,
-0.007450980134308338,
0.11999548226594925,
-0.03035135380923748,
0.04821372777223587,
0.13066574931144714,
0.0440254770219326,
-0.10783413052558899,
0.033060941845178604,
0.034991905093193054,
-0.039596833288669586,
-0.19061638414859772,
-0.01955413445830345,
-0.0460742823779583,
0.017535662278532982,
0.09903786331415176,
0.028780780732631683,
0.028668979182839394,
0.07774002104997635,
0.01618318259716034,
0.0749795064330101,
-0.013391866348683834,
0.07809770852327347,
0.08251409977674484,
0.03078843653202057,
0.1284647285938263,
-0.04230022057890892,
-0.0762445256114006,
0.03427186235785484,
0.006322118919342756,
0.19180266559123993,
0.014034448191523552,
0.13520950078964233,
0.03476211428642273,
0.12321159243583679,
0.0004752769600600004,
0.07078728824853897,
-0.022774899378418922,
-0.0613274946808815,
-0.01822027936577797,
-0.04788653180003166,
-0.03860939294099808,
0.04742494970560074,
-0.08002045005559921,
0.060635536909103394,
-0.11174392700195312,
0.004209819249808788,
0.0585501603782177,
0.20366522669792175,
0.07101991772651672,
-0.332301527261734,
-0.09497977793216705,
0.043346941471099854,
-0.023090140894055367,
-0.04221276938915253,
0.023555345833301544,
0.15099522471427917,
-0.06256665289402008,
0.028166187927126884,
-0.08076190203428268,
0.08282388746738434,
-0.048974744975566864,
0.04884851351380348,
0.015509270131587982,
0.07094719260931015,
-0.029779542237520218,
0.0658411830663681,
-0.2820812165737152,
0.2698727250099182,
0.017548037692904472,
0.06852511316537857,
-0.05218171328306198,
-0.006727451924234629,
0.019683316349983215,
0.07154643535614014,
0.08007698506116867,
-0.019174614921212196,
-0.06635785847902298,
-0.20584164559841156,
-0.06162617355585098,
0.03168439120054245,
0.1145758256316185,
-0.05355197563767433,
0.1259472817182541,
-0.026929795742034912,
0.005405012983828783,
0.07478143274784088,
-0.004231990315020084,
-0.06837648898363113,
-0.10214442014694214,
0.00042335569742135704,
0.04663044214248657,
-0.012622032314538956,
-0.08097966015338898,
-0.09043596684932709,
-0.10882281512022018,
0.16466540098190308,
-0.028280185535550117,
-0.03784799948334694,
-0.10997822880744934,
0.07169009000062943,
0.05847747251391411,
-0.08087479323148727,
0.027680590748786926,
0.005892216227948666,
0.09244886040687561,
0.012703021988272667,
-0.042273227125406265,
0.12962622940540314,
-0.06881218403577805,
-0.17801514267921448,
-0.06719785928726196,
0.12777580320835114,
0.017236217856407166,
0.048334330320358276,
-0.008152570575475693,
0.022303560748696327,
-0.029821021482348442,
-0.07412173599004745,
0.04450938105583191,
-0.02333180047571659,
0.04867929220199585,
0.02005879208445549,
-0.03728475421667099,
0.017071040347218513,
-0.061158619821071625,
-0.03479014337062836,
0.17449389398097992,
0.3027888238430023,
-0.06317433714866638,
-0.0021706214174628258,
0.0538809709250927,
-0.05841289833188057,
-0.18932707607746124,
0.04322276636958122,
0.02474154345691204,
0.013056356459856033,
0.043881941586732864,
-0.15034575760364532,
0.08141390979290009,
0.08668991923332214,
-0.020030032843351364,
0.13852091133594513,
-0.2994435131549835,
-0.1350344568490982,
0.11653319001197815,
0.16290093958377838,
0.14127671718597412,
-0.17856617271900177,
-0.04245128482580185,
-0.038515977561473846,
-0.10947372764348984,
0.08626848459243774,
-0.11906243115663528,
0.12281192094087601,
-0.006228633224964142,
0.07391258329153061,
0.020707063376903534,
-0.06298647075891495,
0.13562025129795074,
-0.025844888761639595,
0.10025914013385773,
-0.0748676061630249,
-0.013289213180541992,
0.048625607043504715,
-0.06204492971301079,
0.02580965869128704,
-0.11932065337896347,
0.02925686351954937,
-0.07712332904338837,
-0.03908400610089302,
-0.05335284024477005,
0.022949840873479843,
-0.030722852796316147,
-0.06977390497922897,
-0.03472740948200226,
0.023095345124602318,
0.04084407538175583,
0.001053451793268323,
0.15470702946186066,
-0.019116898998618126,
0.17401115596294403,
0.1313970386981964,
0.08078134059906006,
-0.06621355563402176,
0.004470796789973974,
0.004013949539512396,
-0.02744334749877453,
0.05056840926408768,
-0.1559404879808426,
0.03410866856575012,
0.12411126494407654,
0.001923451665788889,
0.15638791024684906,
0.07506560534238815,
-0.04209553822875023,
0.027270350605249405,
0.07291191816329956,
-0.18649715185165405,
-0.12543395161628723,
-0.027403367683291435,
-0.02631588838994503,
-0.12558864057064056,
0.05546681955456734,
0.1413574516773224,
-0.06549706310033798,
0.008662586100399494,
-0.01937008649110794,
0.01836203597486019,
-0.03226645663380623,
0.181874617934227,
0.060097310692071915,
0.05310848727822304,
-0.07163658738136292,
0.06028835102915764,
0.0415005087852478,
-0.07720165699720383,
0.03296877443790436,
0.0714295506477356,
-0.07836786657571793,
-0.046359289437532425,
0.03148436173796654,
0.19981014728546143,
-0.0299381110817194,
-0.034338612109422684,
-0.158227801322937,
-0.11072712391614914,
0.047966089099645615,
0.1661681830883026,
0.07752784341573715,
0.004803935531526804,
-0.029686804860830307,
0.019596705213189125,
-0.12300241738557816,
0.12047132104635239,
0.044097598642110825,
0.08863075822591782,
-0.14205990731716156,
0.13484299182891846,
-0.016216667369008064,
0.008660328574478626,
-0.02954987809062004,
0.03689657524228096,
-0.12350223958492279,
0.0024885181337594986,
-0.13360901176929474,
-0.021141543984413147,
-0.022508108988404274,
-0.012571600265800953,
-0.005102424416691065,
-0.05999813601374626,
-0.06017051637172699,
-0.0002872564655262977,
-0.10170474648475647,
-0.023649688810110092,
0.021651310846209526,
0.03288932517170906,
-0.11456331610679626,
-0.038618218153715134,
0.02689896710216999,
-0.07551847398281097,
0.07651444524526596,
0.02184714563190937,
0.016641056165099144,
0.050109270960092545,
-0.17801760137081146,
0.045374900102615356,
0.0563981756567955,
-0.002968628890812397,
0.03729653358459473,
-0.07667995989322662,
-0.01737712323665619,
-0.014389334246516228,
0.060993414372205734,
0.027813974767923355,
0.08757786452770233,
-0.12123914062976837,
0.01214001514017582,
-0.021203039214015007,
-0.06886055320501328,
-0.057615894824266434,
0.03440406545996666,
0.0508849211037159,
0.008411191403865814,
0.17912748456001282,
-0.0995754525065422,
0.03140265494585037,
-0.20807930827140808,
0.005052846856415272,
0.010417158715426922,
-0.12310588359832764,
-0.0733405351638794,
-0.060608841478824615,
0.06278857588768005,
-0.05824673920869827,
0.12476034462451935,
-0.0006720881792716682,
0.032318346202373505,
0.033216848969459534,
-0.03807970881462097,
0.061557088047266006,
0.02322242595255375,
0.2213064730167389,
0.038723189383745193,
-0.058016084134578705,
0.012847879901528358,
0.04918511211872101,
0.12425840646028519,
0.09566701203584671,
0.18727096915245056,
0.14636297523975372,
-0.04428952932357788,
0.10825259983539581,
0.025751717388629913,
-0.038268230855464935,
-0.13780976831912994,
0.03842709958553314,
-0.050151802599430084,
0.08904226869344711,
-0.02044510655105114,
0.20658588409423828,
0.12501105666160583,
-0.14784903824329376,
0.006454102229326963,
-0.05960565432906151,
-0.08220015466213226,
-0.1043345257639885,
-0.06277936697006226,
-0.101381815969944,
-0.1368098258972168,
-0.0023676420096307993,
-0.12221664190292358,
0.016141818836331367,
0.10176494717597961,
0.012562226504087448,
-0.014719563536345959,
0.2133667767047882,
0.03457332402467728,
0.024195678532123566,
0.061356984078884125,
0.0025316718965768814,
-0.021007295697927475,
-0.0690581202507019,
-0.08397487550973892,
0.007199120242148638,
-0.01457947213202715,
0.03362729400396347,
-0.04236135259270668,
-0.03944501653313637,
0.03396294638514519,
-0.010646170005202293,
-0.11749374121427536,
-0.003860791679471731,
0.03842722997069359,
0.0536966547369957,
0.042166683822870255,
0.009691697545349598,
0.003609210019931197,
-0.014162899926304817,
0.23563769459724426,
-0.07916171103715897,
-0.05983785167336464,
-0.09544473886489868,
0.2149474322795868,
0.016427673399448395,
0.01609298586845398,
-0.01235334761440754,
-0.10037307441234589,
0.03652966767549515,
0.2321343570947647,
0.18337225914001465,
-0.1017509177327156,
0.003963876515626907,
-0.011363760568201542,
-0.0002726927923504263,
-0.021487800404429436,
0.08731910586357117,
0.10901668667793274,
0.04123995080590248,
-0.08384669572114944,
-0.033155277371406555,
-0.03128406032919884,
-0.00958728976547718,
-0.0400511808693409,
0.06408073008060455,
0.0502309650182724,
0.03415960818529129,
-0.057643357664346695,
0.0633983388543129,
-0.04802964627742767,
-0.09838758409023285,
0.011644395999610424,
-0.2138635516166687,
-0.14297093451023102,
-0.020956259220838547,
0.10245179384946823,
-0.01470573153346777,
0.06515579670667648,
-0.03406499698758125,
-0.0019375832052901387,
0.04683804139494896,
-0.018531622365117073,
-0.05238908529281616,
-0.06079012155532837,
0.057488489896059036,
-0.12434457987546921,
0.21248307824134827,
-0.043535344302654266,
0.01471560075879097,
0.13798432052135468,
0.019881967455148697,
-0.07400984317064285,
0.09292802959680557,
0.04985632747411728,
-0.053729765117168427,
0.031015805900096893,
0.12892314791679382,
-0.03646363690495491,
0.10621078312397003,
0.06651374697685242,
-0.14653420448303223,
0.011335925199091434,
-0.051156748086214066,
-0.06366412341594696,
-0.042823974043130875,
-0.0419045090675354,
-0.06120206415653229,
0.13936491310596466,
0.1888885200023651,
-0.04233388602733612,
0.009337078779935837,
-0.04398851841688156,
0.02142542041838169,
0.07030681520700455,
0.06699121743440628,
-0.030466753989458084,
-0.25560885667800903,
0.008277942426502705,
0.0907432958483696,
-0.014476378448307514,
-0.3476620018482208,
-0.08055384457111359,
-0.017703033983707428,
-0.02968006208539009,
-0.08987921476364136,
0.09103437513113022,
0.12501169741153717,
0.05133146047592163,
-0.06346820294857025,
-0.09493578225374222,
-0.06787610799074173,
0.17086674273014069,
-0.1173391118645668,
-0.10081101208925247
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"language": ["en"], "license": "apache-2.0", "library_name": "transformers", "pipeline_tag": "conversational"} | text-generation | apatidar0/chat_style_phi-2 | [
"transformers",
"safetensors",
"phi",
"text-generation",
"conversational",
"custom_code",
"en",
"arxiv:1910.09700",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"4-bit",
"region:us"
] | 2024-02-06T08:05:59+00:00 | [
"1910.09700"
] | [
"en"
] | TAGS
#transformers #safetensors #phi #text-generation #conversational #custom_code #en #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #phi #text-generation #conversational #custom_code #en #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
68,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #phi #text-generation #conversational #custom_code #en #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05304252728819847,
0.21067693829536438,
-0.005285065621137619,
0.015017993748188019,
0.09488703310489655,
0.00841852743178606,
0.06399186700582504,
0.11650046706199646,
-0.056802961975336075,
0.1257191300392151,
0.044529084116220474,
0.11682984977960587,
0.1185559555888176,
0.15182296931743622,
-0.009572125971317291,
-0.21556279063224792,
0.049783479422330856,
-0.09544818103313446,
-0.010143846273422241,
0.12479330599308014,
0.15112926065921783,
-0.10353261977434158,
0.06991244852542877,
-0.02506031095981598,
-0.0291392020881176,
-0.035314299166202545,
-0.054662950336933136,
-0.046247292309999466,
0.04020877927541733,
0.04499981552362442,
0.07202181965112686,
0.0030272139701992273,
0.08998805284500122,
-0.29758185148239136,
0.014973890967667103,
0.0653725415468216,
-0.001191635150462389,
0.0684969574213028,
0.08399958908557892,
-0.0649813711643219,
0.11168891936540604,
-0.054221007972955704,
0.14141741394996643,
0.07894281297922134,
-0.09047819674015045,
-0.17373381555080414,
-0.08211427181959152,
0.12009045481681824,
0.17577499151229858,
0.055574871599674225,
-0.03280756622552872,
0.1061437726020813,
-0.08035188168287277,
0.018445637077093124,
0.04377962276339531,
-0.10661302506923676,
-0.06123536825180054,
0.09032029658555984,
0.10073718428611755,
0.052627675235271454,
-0.12326540052890778,
-0.02921311743557453,
0.019688446074724197,
0.02152267098426819,
0.08400408923625946,
0.01732030138373375,
0.14728030562400818,
0.03702453151345253,
-0.13533906638622284,
-0.0700196698307991,
0.10921754688024521,
0.03815252706408501,
-0.0358470119535923,
-0.2351922243833542,
-0.016144176945090294,
-0.03011850267648697,
-0.03592855855822563,
-0.042126432061195374,
0.044829756021499634,
0.0010957319755107164,
0.0870608389377594,
-0.009688428603112698,
-0.07891664654016495,
-0.04300558194518089,
0.08158588409423828,
0.04591892659664154,
0.027943404391407967,
-0.018175246194005013,
0.024238748475909233,
0.10914776474237442,
0.08512882143259048,
-0.11301886290311813,
-0.04981685429811478,
-0.05992940440773964,
-0.06954992562532425,
-0.0436704084277153,
0.03652353957295418,
0.042080558836460114,
0.07725732773542404,
0.2504896819591522,
0.025668220594525337,
0.049143340438604355,
0.034276846796274185,
0.009505282156169415,
0.048325639218091965,
0.09720462560653687,
-0.05420143902301788,
-0.1276749074459076,
-0.02361486107110977,
0.10134486109018326,
0.007463356014341116,
-0.028024982661008835,
-0.036721859127283096,
0.06448429822921753,
0.04582279548048973,
0.1094309538602829,
0.090780109167099,
0.020940512418746948,
-0.07614424079656601,
-0.048730362206697464,
0.19055090844631195,
-0.15357227623462677,
0.03630972281098366,
0.035233598202466965,
-0.03130330145359039,
-0.043374985456466675,
0.00797354057431221,
0.037586044520139694,
-0.036694902926683426,
0.09073301404714584,
-0.05529140681028366,
-0.05345293506979942,
-0.1119932234287262,
-0.031950127333402634,
0.04369383305311203,
0.010677767917513847,
-0.029236091300845146,
-0.030533695593476295,
-0.09382390230894089,
-0.0843011811375618,
0.0917493999004364,
-0.06420957297086716,
-0.07392169535160065,
-0.03168788552284241,
-0.07769164443016052,
0.017875071614980698,
0.018821682780981064,
0.08788267523050308,
-0.02743343822658062,
0.052201781421899796,
-0.05699770152568817,
0.04504300653934479,
0.10616666823625565,
0.03576366975903511,
-0.07105077058076859,
0.07314669340848923,
-0.2061103880405426,
0.0858243778347969,
-0.07866466045379639,
0.04649568349123001,
-0.1619405597448349,
-0.027528908103704453,
0.03458024561405182,
0.01812685839831829,
-0.002447956008836627,
0.13004055619239807,
-0.19379477202892303,
-0.017440907657146454,
0.17665526270866394,
-0.10286331176757812,
-0.08255616575479507,
0.055985670536756516,
-0.05623898282647133,
0.11561261117458344,
0.03521088510751724,
0.020902395248413086,
0.05437091365456581,
-0.10596124082803726,
-0.012196166440844536,
-0.051635827869176865,
-0.007560438010841608,
0.1239284798502922,
0.0804530456662178,
-0.09214013814926147,
0.03855635225772858,
0.018793022260069847,
-0.03651098534464836,
-0.06763613969087601,
-0.031990550458431244,
-0.10343429446220398,
0.011474592611193657,
-0.08307518064975739,
0.008785050362348557,
-0.01107749156653881,
-0.09670176357030869,
-0.031102923676371574,
-0.15749146044254303,
-0.01946769282221794,
0.08737806230783463,
-0.007719618733972311,
-0.024884242564439774,
-0.10628359764814377,
0.03167852759361267,
0.012423324398696423,
-0.009718629531562328,
-0.1252046674489975,
-0.023749809712171555,
0.03154051676392555,
-0.14699135720729828,
0.02527863346040249,
-0.06713863462209702,
0.04874741658568382,
0.015196514315903187,
-0.03298112377524376,
-0.026186710223555565,
0.013393213972449303,
0.01666373573243618,
-0.027283232659101486,
-0.22932931780815125,
-0.025383027270436287,
-0.033826448023319244,
0.16906790435314178,
-0.2311459481716156,
0.039328400045633316,
0.05890214443206787,
0.1474994271993637,
-0.0020864943508058786,
-0.05547427386045456,
0.024057798087596893,
-0.061536747962236404,
-0.02739359252154827,
-0.053940482437610626,
0.001484034932218492,
-0.012396818026900291,
-0.0417022742331028,
0.019665291532874107,
-0.1721620410680771,
-0.03679148107767105,
0.10256116092205048,
0.04645870253443718,
-0.12329006940126419,
-0.035213373601436615,
-0.03168052062392235,
-0.05534784868359566,
-0.04332776740193367,
-0.058579184114933014,
0.09623181074857712,
0.05637683719396591,
0.04320797696709633,
-0.07009130716323853,
-0.07116712629795074,
-0.0019244886934757233,
-0.02461167611181736,
-0.024296749383211136,
0.09722871333360672,
0.08738895505666733,
-0.12265099585056305,
0.09720544517040253,
0.079180046916008,
0.05824220925569534,
0.08175231516361237,
-0.022626236081123352,
-0.07492133975028992,
-0.026115668937563896,
0.03508786857128143,
0.0211217999458313,
0.1328592598438263,
-0.06888367235660553,
0.042247481644153595,
0.04469481483101845,
-0.03296691179275513,
0.024794667959213257,
-0.08755132555961609,
0.017609786242246628,
0.024319984018802643,
-0.022299764677882195,
0.02867230214178562,
-0.04294005408883095,
0.01292547956109047,
0.08620108664035797,
0.04952329397201538,
0.021203964948654175,
0.01934974081814289,
-0.05039916932582855,
-0.11553249508142471,
0.1617971509695053,
-0.11480218172073364,
-0.20416958630084991,
-0.13359209895133972,
0.02955988235771656,
0.037268608808517456,
-0.015322685241699219,
-0.00143814564216882,
-0.0530034638941288,
-0.10281530767679214,
-0.0938793197274208,
0.00529107078909874,
0.0395987443625927,
-0.09064576774835587,
-0.044990573078393936,
0.037639498710632324,
0.0428447350859642,
-0.1384909301996231,
0.014739788137376308,
0.04441283643245697,
-0.08553542196750641,
-0.010927818715572357,
0.058337874710559845,
0.0896209254860878,
0.19911271333694458,
0.013011621311306953,
-0.010525044053792953,
0.024125995114445686,
0.2276141345500946,
-0.13670286536216736,
0.09853033721446991,
0.13383691012859344,
-0.07349269092082977,
0.08091839402914047,
0.21293212473392487,
0.04146874323487282,
-0.09347618371248245,
0.023768894374370575,
0.039225198328495026,
-0.02451152727007866,
-0.2455354630947113,
-0.0724734216928482,
-0.0035798652097582817,
-0.05802377313375473,
0.08359584212303162,
0.08792316913604736,
0.10060075670480728,
0.03711729496717453,
-0.08443194627761841,
-0.08768611401319504,
0.059964753687381744,
0.10911842435598373,
-0.006800605449825525,
0.006196596659719944,
0.08849874138832092,
-0.0339326448738575,
0.01866334304213524,
0.08908650279045105,
0.01933460682630539,
0.1419389396905899,
0.0477035827934742,
0.17215991020202637,
0.09121091663837433,
0.08508290350437164,
0.002500757807865739,
0.02159143052995205,
0.008237913250923157,
0.04367424175143242,
-0.00016885984223335981,
-0.07858772575855255,
-0.019606517627835274,
0.11813430488109589,
0.0542515330016613,
0.018860533833503723,
0.015826400369405746,
-0.039203446358442307,
0.07557416707277298,
0.19802245497703552,
0.00041787122609093785,
-0.20252496004104614,
-0.05511176213622093,
0.07728883624076843,
-0.0911545380949974,
-0.10664982348680496,
-0.0008576199761591852,
0.022280288860201836,
-0.16762295365333557,
0.0399266853928566,
-0.0297061949968338,
0.10913378745317459,
-0.11270272731781006,
-0.02000441960990429,
0.07450654357671738,
0.062085941433906555,
-0.01539277657866478,
0.07085306942462921,
-0.20508484542369843,
0.11064594984054565,
0.009464170783758163,
0.07356847822666168,
-0.0982123464345932,
0.08668678253889084,
-0.0055372221395373344,
-0.02098574861884117,
0.15836907923221588,
-0.006478649098426104,
-0.06921280175447464,
-0.0692707821726799,
-0.09336661547422409,
-0.006852633319795132,
0.09906984865665436,
-0.1293259859085083,
0.08427618443965912,
-0.032012175768613815,
-0.031582802534103394,
-0.0006655828910879791,
-0.09964590519666672,
-0.11439529061317444,
-0.17176322638988495,
0.049322690814733505,
-0.12063746154308319,
0.036295194178819656,
-0.10895982384681702,
-0.02689971588551998,
-0.03555702790617943,
0.18198469281196594,
-0.19973859190940857,
-0.08154378831386566,
-0.1370941400527954,
-0.09625259041786194,
0.13574929535388947,
-0.04112908989191055,
0.0988810732960701,
-0.011401891708374023,
0.1624232679605484,
0.010679596103727818,
-0.011911964975297451,
0.07823657989501953,
-0.09174399077892303,
-0.20643159747123718,
-0.06686526536941528,
0.1633608639240265,
0.10926523059606552,
0.03745380416512489,
0.004766597878187895,
0.03538455069065094,
-0.02709176018834114,
-0.10998202115297318,
0.025372494012117386,
0.1384047418832779,
0.08278309553861618,
0.003361782059073448,
-0.015090616419911385,
-0.1254016011953354,
-0.08132994920015335,
-0.044294748455286026,
0.02253984659910202,
0.16261626780033112,
-0.07484741508960724,
0.15310229361057281,
0.13581934571266174,
-0.06390982866287231,
-0.201278418302536,
0.004602800123393536,
0.023177826777100563,
-0.00892256386578083,
0.010775944218039513,
-0.17851300537586212,
0.0823143869638443,
0.009632361121475697,
-0.057672567665576935,
0.08335928618907928,
-0.18745218217372894,
-0.13742350041866302,
0.08357598632574081,
0.056024376302957535,
-0.2059844732284546,
-0.13938885927200317,
-0.09446496516466141,
-0.040920745581388474,
-0.1596170961856842,
0.08976589143276215,
-0.0021923910826444626,
0.00197130860760808,
0.036902375519275665,
0.01170450821518898,
0.02716280333697796,
-0.05522473156452179,
0.18292078375816345,
-0.0006855015526525676,
0.03169224038720131,
-0.08406304568052292,
-0.09826512634754181,
0.02778751403093338,
-0.04561195522546768,
0.07611318677663803,
-0.03207550197839737,
0.015659049153327942,
-0.11383643001317978,
-0.042982764542102814,
-0.056322306394577026,
0.013263056054711342,
-0.10127153247594833,
-0.09482406824827194,
-0.049248017370700836,
0.08554179221391678,
0.10633818060159683,
-0.018857521936297417,
-0.04034579545259476,
-0.0826563909649849,
0.06565643101930618,
0.22093859314918518,
0.19312402606010437,
0.0792030617594719,
-0.06796590238809586,
0.002076240722090006,
-0.026919696480035782,
0.04416809603571892,
-0.20036627352237701,
0.05431428551673889,
0.059862490743398666,
0.019700434058904648,
0.10984740406274796,
-0.025107795372605324,
-0.1478133350610733,
-0.07218055427074432,
0.06228494271636009,
-0.06460188329219818,
-0.20433777570724487,
0.01477841380983591,
0.05293165519833565,
-0.16930407285690308,
-0.040359977632761,
0.033246200531721115,
-0.015458638779819012,
-0.034837573766708374,
0.009771597571671009,
0.09310673177242279,
-0.010328204371035099,
0.08933914452791214,
0.07675506919622421,
0.09257932752370834,
-0.09725937247276306,
0.08652380853891373,
0.09892547875642776,
-0.06084885448217392,
0.02920508198440075,
0.08536601811647415,
-0.0522179901599884,
-0.036240044981241226,
0.05553547665476799,
0.09959111362695694,
0.01999804936349392,
-0.059811897575855255,
0.00006413876690203324,
-0.09504769742488861,
0.06052432209253311,
0.10880861431360245,
0.025225933641195297,
0.014429938979446888,
0.05478491634130478,
0.03475836291909218,
-0.08852169662714005,
0.12083707004785538,
0.05268307402729988,
0.013723455369472504,
-0.04565272480249405,
-0.011821814812719822,
0.010673360899090767,
-0.03306866064667702,
-0.0036077252589166164,
-0.0066661895252764225,
-0.07656672596931458,
-0.005622462369501591,
-0.13715621829032898,
0.014399672858417034,
-0.07747819274663925,
0.012098930776119232,
0.027033595368266106,
-0.02339104190468788,
0.006096499506384134,
-0.0011639948934316635,
-0.07559714466333389,
-0.05561721697449684,
-0.007243456318974495,
0.10027989000082016,
-0.16574880480766296,
0.021191151812672615,
0.08179901540279388,
-0.10774249583482742,
0.09212930500507355,
-0.008173361420631409,
-0.006225619465112686,
0.008181573823094368,
-0.1536961793899536,
0.05460550636053085,
-0.026331916451454163,
0.007686531636863947,
0.0055334921926259995,
-0.17470014095306396,
-0.002055919263511896,
-0.03106197714805603,
-0.07348819077014923,
-0.007093891967087984,
-0.0230522770434618,
-0.10993257910013199,
0.0991521030664444,
0.012704988941550255,
-0.08263786882162094,
-0.023490095511078835,
0.03949439898133278,
0.08573287725448608,
-0.040069419890642166,
0.14250704646110535,
-0.01709291711449623,
0.061692606657743454,
-0.1690875142812729,
-0.010849925689399242,
-0.010512018576264381,
0.019159989431500435,
-0.05858876556158066,
-0.006176783237606287,
0.05015026405453682,
-0.020747967064380646,
0.1919647604227066,
-0.02485174499452114,
0.01107171643525362,
0.05921786651015282,
0.03064834140241146,
0.006688857916742563,
0.10017140209674835,
0.06863027811050415,
0.010215438902378082,
0.007950753904879093,
0.011409591883420944,
-0.04994957894086838,
-0.03604850172996521,
-0.1736983358860016,
0.06355947256088257,
0.20414267480373383,
0.10474557429552078,
-0.02166396751999855,
0.06958560645580292,
-0.11506994813680649,
-0.10128288716077805,
0.13099665939807892,
-0.03928308188915253,
-0.0007606901926919818,
-0.07613517343997955,
0.14544852077960968,
0.14271768927574158,
-0.19256561994552612,
0.07072252035140991,
-0.07632597535848618,
-0.04810631275177002,
-0.10403475165367126,
-0.20319993793964386,
-0.064572274684906,
-0.04172450304031372,
-0.012059156782925129,
-0.0554678738117218,
0.06602314859628677,
0.09152829647064209,
-0.0046806721948087215,
-0.016190307214856148,
0.0714721828699112,
-0.03400355949997902,
0.004607755690813065,
0.02896539308130741,
0.05756030231714249,
0.017324240878224373,
-0.06504013389348984,
0.009229636751115322,
-0.012860128656029701,
0.04654857888817787,
0.07784947007894516,
0.037033889442682266,
-0.02663588896393776,
0.016296256333589554,
-0.026662074029445648,
-0.10297907888889313,
0.04998348280787468,
-0.02642800100147724,
-0.05538840591907501,
0.15183469653129578,
0.02420620433986187,
0.0066150035709142685,
-0.00979364663362503,
0.2206345647573471,
-0.0663769543170929,
-0.10395718365907669,
-0.1495487093925476,
0.08193647861480713,
-0.05330151692032814,
0.04774538427591324,
0.05393621325492859,
-0.11190299689769745,
0.02671905606985092,
0.15964439511299133,
0.16352178156375885,
-0.03423239663243294,
0.013837534934282303,
0.03136952593922615,
0.005134724546223879,
-0.029351139441132545,
0.04160432517528534,
0.05016618221998215,
0.13908687233924866,
-0.06360635906457901,
0.07120645046234131,
0.0045278025791049,
-0.08282969892024994,
-0.022290311753749847,
0.1260528862476349,
-0.012291675433516502,
0.0022220194805413485,
-0.05382151901721954,
0.12001240998506546,
-0.07085191458463669,
-0.20803196728229523,
0.04152848944067955,
-0.07544092833995819,
-0.13136893510818481,
-0.02671942114830017,
0.043720412999391556,
-0.001413466059602797,
0.019457625225186348,
0.07242117077112198,
-0.06774432212114334,
0.17095641791820526,
0.03499859198927879,
-0.06565544754266739,
-0.05062325298786163,
0.08196157962083817,
-0.07826371490955353,
0.30787762999534607,
0.018345927819609642,
0.047443270683288574,
0.10794367641210556,
-0.01775423064827919,
-0.12261483818292618,
0.03012436255812645,
0.10430839657783508,
-0.07217620313167572,
0.05330660939216614,
0.15697994828224182,
-0.004371391609311104,
0.12886188924312592,
0.07006407529115677,
-0.07786815613508224,
0.04820581525564194,
-0.07887857407331467,
-0.0750528946518898,
-0.10475457459688187,
0.09406612813472748,
-0.08905436843633652,
0.14869755506515503,
0.12406081706285477,
-0.05842268466949463,
0.011233458295464516,
-0.02572404406964779,
0.0612497441470623,
-0.00620219437405467,
0.12421134114265442,
0.014326926320791245,
-0.18744800984859467,
0.030852284282445908,
-0.03156835958361626,
0.10030072182416916,
-0.18521787226200104,
-0.07952380180358887,
0.04049504175782204,
0.001009613974019885,
-0.08310295641422272,
0.11888335645198822,
0.07643375545740128,
0.030144063755869865,
-0.05040854588150978,
-0.01831655576825142,
-0.013163668103516102,
0.1492988020181656,
-0.09344146400690079,
-0.008071952499449253
] |
null | null | transformers | Thanks to @Epiculous for the dope model/ help with llm backends and support overall.
Id like to also thank @kalomaze for the dope sampler additions to ST.
@SanjiWatsuki Thank you very much for the help, and the model!
ST users can find the TextGenPreset in the folder labeled so.

Quants:Thank you @bartowski, @jeiku, @konz00.
https://huggingface.co/bartowski/Kunocchini-exl2
https://huggingface.co/jeiku/Konocchini-7B_GGUF
https://huggingface.co/konz00/Kunocchini-7b-GGUF
The following models were included in the merge:
* [SanjiWatsuki/Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
* [Epiculous/Fett-uccine-7B](https://huggingface.co/Epiculous/Fett-uccine-7B)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
slices:
- sources:
- model: SanjiWatsuki/Kunoichi-DPO-v2-7B
layer_range: [0, 32]
- model: Epiculous/Fett-uccine-7B
layer_range: [0, 32]
merge_method: slerp
base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5
dtype: bfloat16
``` | {"license": "other", "library_name": "transformers", "tags": ["mergekit", "merge", "alpaca", "mistral"], "base_model": ["SanjiWatsuki/Kunoichi-DPO-v2-7B", "Epiculous/Fett-uccine-7B"]} | text-generation | Test157t/Kunocchini-7b | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"alpaca",
"base_model:SanjiWatsuki/Kunoichi-DPO-v2-7B",
"base_model:Epiculous/Fett-uccine-7B",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T08:08:29+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #mergekit #merge #alpaca #base_model-SanjiWatsuki/Kunoichi-DPO-v2-7B #base_model-Epiculous/Fett-uccine-7B #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Thanks to @Epiculous for the dope model/ help with llm backends and support overall.
Id like to also thank @kalomaze for the dope sampler additions to ST.
@SanjiWatsuki Thank you very much for the help, and the model!
ST users can find the TextGenPreset in the folder labeled so.
!image/jpeg
Quants:Thank you @bartowski, @jeiku, @konz00.
URL
URL
URL
The following models were included in the merge:
* SanjiWatsuki/Kunoichi-DPO-v2-7B
* Epiculous/Fett-uccine-7B
### Configuration
The following YAML configuration was used to produce this model:
| [
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #alpaca #base_model-SanjiWatsuki/Kunoichi-DPO-v2-7B #base_model-Epiculous/Fett-uccine-7B #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
101,
17
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #alpaca #base_model-SanjiWatsuki/Kunoichi-DPO-v2-7B #base_model-Epiculous/Fett-uccine-7B #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Configuration\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.04827842116355896,
-0.02711334638297558,
-0.0018630074337124825,
0.01664290390908718,
0.11920790374279022,
0.03723376244306564,
0.21972453594207764,
0.06883828341960907,
0.007329242303967476,
-0.00759770767763257,
0.09230688214302063,
0.08492182940244675,
0.047438375651836395,
0.13973461091518402,
-0.03369012475013733,
-0.22017674148082733,
0.07378402352333069,
0.025632141157984734,
-0.07307296246290207,
0.07651963829994202,
0.12004659324884415,
-0.04043719545006752,
0.1106174886226654,
0.013773120939731598,
-0.17258219420909882,
0.05627387389540672,
-0.05822443589568138,
-0.042880889028310776,
0.0683760717511177,
0.12694860994815826,
0.09840621799230576,
0.10027120262384415,
0.040738143026828766,
-0.12937432527542114,
0.035205185413360596,
-0.028859740123152733,
-0.06325878202915192,
0.010682976804673672,
0.0640692189335823,
-0.008216544054448605,
0.07740597426891327,
0.036575157195329666,
-0.054907478392124176,
-0.007223848253488541,
-0.09200245141983032,
0.0019574244506657124,
-0.025775820016860962,
0.13029859960079193,
0.11649122089147568,
0.04255615174770355,
-0.0029746396467089653,
-0.009635659866034985,
-0.03555180877447128,
0.029610145837068558,
0.05114969611167908,
-0.26973220705986023,
-0.014907448552548885,
0.10466357320547104,
0.0672745332121849,
-0.045779939740896225,
0.04325011000037193,
0.02393496036529541,
0.0937284380197525,
-0.054937608540058136,
-0.033725637942552567,
-0.04857064411044121,
0.21965603530406952,
-0.006071420852094889,
-0.0855836495757103,
-0.00603888463228941,
0.24262064695358276,
-0.003680033143609762,
-0.007015388458967209,
-0.02931125834584236,
-0.08313493430614471,
0.10225141048431396,
-0.04623621702194214,
0.010182720609009266,
-0.008013308048248291,
0.061975058168172836,
0.09878524392843246,
-0.009275433607399464,
-0.10985403507947922,
-0.03837134316563606,
-0.13768935203552246,
0.1961567997932434,
0.05391858518123627,
0.02191370353102684,
-0.08207295835018158,
0.08884197473526001,
-0.08827128261327744,
-0.14191555976867676,
0.019835561513900757,
-0.08384633809328079,
0.06798641383647919,
-0.013442869298160076,
-0.05041256174445152,
-0.054055117070674896,
0.150286003947258,
0.14547504484653473,
0.032400477677583694,
-0.008257316425442696,
0.06960216164588928,
0.05128570646047592,
0.02375795878469944,
0.025884009897708893,
-0.152353435754776,
-0.0916401818394661,
0.04324277117848396,
0.027787096798419952,
0.0602242611348629,
0.0006381332059390843,
-0.2061772495508194,
-0.04940468817949295,
-0.0019161923555657268,
0.021098384633660316,
0.040300458669662476,
0.11528203636407852,
-0.03900331258773804,
-0.049359697848558426,
0.15175774693489075,
-0.07664788514375687,
0.006479327566921711,
-0.006913101300597191,
0.0018476698314771056,
-0.019524691626429558,
0.024962669238448143,
0.06022161990404129,
-0.0022848027292639017,
0.08436019718647003,
-0.057056721299886703,
-0.032272566109895706,
-0.08938269317150116,
-0.05505083128809929,
0.016590017825365067,
-0.04091981053352356,
0.07154738903045654,
-0.11250369995832443,
-0.2414228916168213,
-0.00477973697707057,
0.06559791415929794,
-0.04949072375893593,
-0.04507135972380638,
-0.04922318458557129,
0.02954636514186859,
0.007037205155938864,
-0.05498412624001503,
-0.015010728500783443,
-0.051807086914777756,
-0.0064636170864105225,
0.009084993973374367,
0.02506115287542343,
-0.17610737681388855,
0.06116442754864693,
-0.07826689630746841,
0.08544585853815079,
-0.06735920161008835,
0.0417759045958519,
-0.0010660379193723202,
0.13097339868545532,
-0.02672869712114334,
0.027832165360450745,
-0.06682957708835602,
0.05708111450076103,
0.03392736613750458,
0.17229638993740082,
-0.0823526382446289,
-0.08085040748119354,
0.0953034833073616,
-0.130548357963562,
-0.1739717572927475,
0.06149088591337204,
-0.009196320548653603,
0.09962261468172073,
0.042467594146728516,
0.19154810905456543,
0.06219761446118355,
-0.03484632074832916,
0.0376192182302475,
0.02101972885429859,
-0.04239918664097786,
-0.10335312783718109,
0.06743117421865463,
0.026657341048121452,
-0.1604180485010147,
0.06812530010938644,
0.010953648947179317,
0.08808039128780365,
-0.03727931156754494,
-0.051956404000520706,
-0.06657702475786209,
-0.06575150042772293,
-0.02463393285870552,
-0.014593122527003288,
0.06325376778841019,
-0.02295079454779625,
-0.006074232514947653,
0.10872043669223785,
0.06880102306604385,
-0.06844455003738403,
-0.0037995390594005585,
-0.054548367857933044,
0.1602417379617691,
-0.07685508579015732,
0.046639710664749146,
-0.05682354420423508,
0.019746629521250725,
-0.01147641334682703,
0.0781083032488823,
0.00909812469035387,
0.0301674772053957,
0.038973353803157806,
0.024747004732489586,
-0.045214708894491196,
-0.01730010099709034,
0.15292508900165558,
0.045303937047719955,
-0.020870434120297432,
-0.1821645051240921,
0.006694034673273563,
-0.05939997360110283,
0.15425150096416473,
-0.06110251694917679,
0.047159191220998764,
-0.04149899259209633,
0.1758696585893631,
-0.0697193518280983,
0.11764055490493774,
0.02803376503288746,
0.0361005999147892,
-0.06132164224982262,
0.01783384010195732,
0.09938912093639374,
0.03164348378777504,
-0.10619216412305832,
0.16097582876682281,
-0.1391427367925644,
0.21448422968387604,
0.14588069915771484,
-0.01505084615200758,
0.029427533969283104,
-0.13240383565425873,
-0.010726837441325188,
-0.03490781411528587,
0.06956464797258377,
-0.006944119930267334,
0.0065021757036447525,
-0.05230698361992836,
0.13956686854362488,
-0.09676801413297653,
0.03225577995181084,
-0.006882560905069113,
-0.07261849194765091,
-0.052757203578948975,
0.07924531400203705,
0.05450797080993652,
-0.22328446805477142,
0.16451714932918549,
0.18089906871318817,
0.014357209205627441,
0.18721279501914978,
-0.024304062128067017,
-0.01722545735538006,
-0.013235867954790592,
0.03659934550523758,
-0.01709354855120182,
0.019855083897709846,
-0.09953002631664276,
0.012102420441806316,
0.043956585228443146,
-0.04188738390803337,
0.0940956100821495,
-0.1070147305727005,
-0.022475875914096832,
0.03815024718642235,
-0.0030133933760225773,
0.024663783609867096,
0.0711938813328743,
-0.03066862002015114,
0.06641784310340881,
-0.03346731513738632,
-0.08228949457406998,
0.06751552224159241,
0.02680816687643528,
-0.0723862573504448,
0.17499059438705444,
-0.10534969717264175,
-0.19768911600112915,
-0.16472530364990234,
-0.12475664913654327,
-0.10290008038282394,
0.05565076693892479,
0.08311803638935089,
-0.010475697927176952,
-0.028615254908800125,
-0.07046202570199966,
0.020567892119288445,
0.005998512264341116,
0.009634182788431644,
-0.048011746257543564,
0.04223409667611122,
-0.022075841203331947,
-0.06295625865459442,
-0.023301955312490463,
0.00312820915132761,
0.021899636834859848,
0.08299488574266434,
-0.14362794160842896,
0.11527347564697266,
0.12103398889303207,
0.021051324903964996,
0.011968267150223255,
0.00023358204634860158,
0.18170946836471558,
-0.019445421174168587,
0.04302765056490898,
0.14740702509880066,
-0.03859515115618706,
0.05904274433851242,
0.30777058005332947,
0.015019409358501434,
-0.027927566319704056,
0.01495936419814825,
-0.0962192639708519,
-0.060434408485889435,
-0.15977980196475983,
-0.09377571940422058,
-0.09919630736112595,
-0.00986449234187603,
0.04843420162796974,
0.051899317651987076,
0.09431259334087372,
0.1268768608570099,
-0.06498804688453674,
0.08645085990428925,
0.019048627465963364,
0.06320440024137497,
0.1873151659965515,
0.004158942494541407,
0.11092329770326614,
-0.05576765164732933,
-0.10851065069437027,
0.05764055252075195,
0.0839732363820076,
0.12245119363069534,
0.0626525804400444,
0.057070229202508926,
0.0781845822930336,
0.10439913719892502,
0.14278045296669006,
0.10202701389789581,
-0.00944024883210659,
-0.027169793844223022,
-0.050301145762205124,
-0.07327336072921753,
0.025996997952461243,
0.09592052549123764,
-0.06503215432167053,
-0.050580430775880814,
-0.028806647285819054,
-0.051263999193906784,
0.04461659863591194,
0.09178977459669113,
0.0572366788983345,
-0.24687190353870392,
0.020684214308857918,
0.07057778537273407,
0.043527062982320786,
-0.0632253959774971,
-0.015055623836815357,
-0.0519869439303875,
-0.005506051704287529,
0.1440400630235672,
0.02944566123187542,
0.13347747921943665,
0.05293964222073555,
0.0410480760037899,
-0.05886189267039299,
0.03541692718863487,
-0.018535375595092773,
0.10193823277950287,
-0.22615087032318115,
0.2454679161310196,
0.016779882833361626,
-0.006454261019825935,
-0.024616781622171402,
0.025521602481603622,
0.03985495865345001,
0.2654792070388794,
0.06567005813121796,
-0.025089414790272713,
-0.0853268951177597,
-0.009436042048037052,
-0.09907782822847366,
0.042868487536907196,
-0.003085381118580699,
-0.04561910405755043,
0.06334614008665085,
-0.01650981605052948,
-0.023145822808146477,
0.0071685598231852055,
0.12452204525470734,
-0.15051111578941345,
-0.17505843937397003,
0.059135545045137405,
0.04532943293452263,
0.07534235715866089,
-0.04809021204710007,
-0.020900728181004524,
-0.06269833445549011,
0.1614396870136261,
0.026788419112563133,
-0.09869156032800674,
-0.11646857857704163,
-0.05850240960717201,
0.09525078535079956,
-0.054383497685194016,
0.037291835993528366,
-0.05640898644924164,
0.021053459495306015,
-0.04212260618805885,
-0.1763952672481537,
0.07766498625278473,
-0.08618403971195221,
-0.014691151678562164,
-0.03837870433926582,
0.061544355005025864,
-0.08359459042549133,
-0.01109226606786251,
0.04320131242275238,
-0.010309847071766853,
-0.10577825456857681,
-0.07340770959854126,
-0.0647248849272728,
0.09122436493635178,
0.06719329953193665,
0.00020075625798199326,
-0.10885480791330338,
-0.1012924537062645,
0.020856603980064392,
-0.06649336963891983,
0.078319251537323,
0.22061479091644287,
-0.030178898945450783,
0.06837313622236252,
0.20603573322296143,
-0.1116124838590622,
-0.24767433106899261,
-0.17581267654895782,
-0.05426565557718277,
0.06437655538320541,
-0.06449539959430695,
-0.1018032506108284,
0.046013325452804565,
0.069807268679142,
-0.04895111545920372,
-0.002665596315637231,
-0.13559457659721375,
-0.15921181440353394,
0.143476665019989,
0.03975959122180939,
0.3384845554828644,
-0.14577636122703552,
-0.07500344514846802,
-0.12743136286735535,
-0.13310754299163818,
0.026317588984966278,
-0.10849227011203766,
0.0857233926653862,
-0.05038416013121605,
-0.015174556523561478,
-0.0009092761902138591,
-0.03481379523873329,
0.13366574048995972,
-0.029639163985848427,
0.04023674502968788,
-0.13447821140289307,
0.01931266114115715,
0.053546030074357986,
-0.031689513474702835,
0.07519202679395676,
-0.16804367303848267,
0.019312376156449318,
-0.03553248569369316,
-0.08398589491844177,
-0.0004303175664972514,
0.047428444027900696,
-0.01930914632976055,
-0.06686064600944519,
-0.04119794815778732,
0.003094941843301058,
-0.02297907881438732,
-0.015424715355038643,
0.10801471024751663,
-0.050220638513565063,
0.05498524010181427,
0.20862607657909393,
0.0642523318529129,
-0.1635848879814148,
0.12353424727916718,
-0.00035796547308564186,
-0.0733039602637291,
0.06026820093393326,
-0.10361682623624802,
-0.004901808220893145,
0.09861240535974503,
-0.05545944720506668,
0.09965869039297104,
0.04748003929853439,
-0.008351662196218967,
-0.029722798615694046,
0.17194798588752747,
-0.14592130482196808,
-0.12672258913516998,
-0.08742935955524445,
-0.012929934076964855,
0.01089996937662363,
0.04428734630346298,
0.10972648859024048,
-0.05944132059812546,
-0.02948269620537758,
0.013803600333631039,
-0.0272052139043808,
-0.07192768901586533,
0.07991348206996918,
0.04484693706035614,
0.013286398723721504,
-0.09596257656812668,
0.05117220804095268,
0.019814668223261833,
-0.022392015904188156,
-0.033993493765592575,
-0.0018929399084299803,
-0.12932735681533813,
-0.10916581749916077,
-0.010762713849544525,
0.19309309124946594,
-0.15516521036624908,
-0.12981760501861572,
-0.13306327164173126,
-0.14265862107276917,
0.047089286148548126,
0.0624978169798851,
0.08002228289842606,
0.03460729494690895,
0.036258239299058914,
-0.07899819314479828,
-0.03279532492160797,
0.0642322525382042,
0.05945814028382301,
0.06186084449291229,
-0.12277395278215408,
-0.05678664147853851,
-0.036862995475530624,
0.04866143316030502,
-0.07160825282335281,
0.018947076052427292,
-0.15680108964443207,
-0.014868396334350109,
-0.2099728137254715,
0.003228678833693266,
-0.1504136621952057,
-0.031203685328364372,
-0.005326984915882349,
-0.047152042388916016,
-0.04951391741633415,
0.012155413627624512,
-0.07574813067913055,
0.0033609603997319937,
-0.045337412506341934,
0.06475501507520676,
-0.01829170435667038,
-0.043299589306116104,
0.034130971878767014,
-0.045721977949142456,
0.02040882408618927,
0.04670795053243637,
-0.0640890970826149,
-0.0000716974536771886,
-0.1544826477766037,
-0.04135836288332939,
0.030543649569153786,
0.020880255848169327,
0.029347719624638557,
-0.07383836805820465,
0.005155343096703291,
0.12674221396446228,
-0.041768915951251984,
-0.01918124593794346,
0.017370689660310745,
-0.06372563540935516,
-0.021881110966205597,
-0.054176315665245056,
-0.07764279842376709,
-0.007076500449329615,
-0.009998956695199013,
0.1064240112900734,
0.06022554263472557,
0.12020079046487808,
-0.05914854630827904,
-0.012112852185964584,
-0.10795850306749344,
0.0023099244572222233,
-0.00588566018268466,
-0.13245926797389984,
-0.05312955006957054,
-0.14467264711856842,
-0.036685530096292496,
0.0465601421892643,
0.25636768341064453,
0.001967947231605649,
-0.0758497565984726,
-0.0013843959895893931,
0.01864277943968773,
0.16856811940670013,
0.057416848838329315,
0.35542088747024536,
0.04007667675614357,
0.047582462430000305,
-0.12627099454402924,
0.06484385579824448,
0.018572038039565086,
0.0702156350016594,
-0.054485633969306946,
0.06004071608185768,
0.0033510439097881317,
0.08897294104099274,
0.07866328954696655,
0.02404113858938217,
0.013654472306370735,
-0.08172906935214996,
-0.03801092877984047,
0.03420780226588249,
0.0007718786946497858,
0.07712253928184509,
0.22012469172477722,
-0.08276065438985825,
0.005649927072227001,
0.006277555134147406,
-0.04309273138642311,
-0.11326184868812561,
-0.13219714164733887,
-0.13525725901126862,
-0.190224289894104,
0.003807170782238245,
-0.07821667194366455,
-0.035936348140239716,
0.035526081919670105,
0.014330904930830002,
-0.014118494465947151,
0.07346297800540924,
0.10065337270498276,
-0.04305102676153183,
0.03137277811765671,
-0.037438105791807175,
-0.0027471515350043774,
-0.008672461844980717,
-0.0522216372191906,
-0.04014093428850174,
-0.06711146980524063,
-0.06662419438362122,
0.021575741469860077,
0.03944654390215874,
0.06558147817850113,
-0.024786842986941338,
-0.03884341940283775,
-0.021313318982720375,
0.035993948578834534,
0.07458102703094482,
0.05339807644486427,
-0.01592826098203659,
-0.05110785365104675,
0.04146649315953255,
0.10867654532194138,
0.0026529403403401375,
-0.16334743797779083,
-0.07465942203998566,
0.09459660947322845,
0.02346835657954216,
0.06702155619859695,
0.024961421266198158,
-0.020753808319568634,
0.02535661682486534,
0.2165154069662094,
0.2999897301197052,
-0.03764677047729492,
0.008607716299593449,
-0.03432457894086838,
0.018979806452989578,
0.058741070330142975,
0.08903783559799194,
0.01440535020083189,
0.1224108338356018,
-0.06261589378118515,
-0.0000893991527846083,
-0.06586897373199463,
-0.034040845930576324,
-0.12257564067840576,
0.0588960275053978,
0.028289133682847023,
-0.112399160861969,
0.002373314229771495,
0.09721720963716507,
-0.11215326189994812,
0.06971370428800583,
0.009897541254758835,
-0.06430715322494507,
-0.0694926381111145,
-0.05228462442755699,
0.13275741040706635,
0.0058716898784041405,
0.026484830304980278,
-0.10139104723930359,
0.023448947817087173,
0.05905943736433983,
-0.013172407634556293,
-0.13941824436187744,
-0.027598802000284195,
0.03425169736146927,
0.06910466402769089,
0.004929791670292616,
-0.0005358982598409057,
0.04458104074001312,
0.10085394233465195,
0.03807670623064041,
-0.09753921627998352,
0.06951480358839035,
-0.003414396895095706,
-0.021824847906827927,
0.09375450760126114,
-0.08520227670669556,
-0.024428464472293854,
-0.09894958883523941,
0.010730480775237083,
-0.10178132355213165,
0.06641587615013123,
-0.01573844812810421,
-0.05112146958708763,
-0.07971490174531937,
0.07574368268251419,
-0.08062592148780823,
0.11270017921924591,
0.07156902551651001,
-0.03344453126192093,
-0.02930723875761032,
-0.006025233305990696,
0.0819372609257698,
0.0334857814013958,
0.024256836622953415,
-0.0012008666526526213,
-0.1393985003232956,
-0.04714465141296387,
0.07845287770032883,
0.03150477260351181,
-0.2207636535167694,
-0.052143797278404236,
-0.14481964707374573,
-0.015454757027328014,
-0.11846723407506943,
0.10451634228229523,
0.20558148622512817,
0.023083413019776344,
-0.03188784793019295,
-0.16700734198093414,
-0.018472736701369286,
0.014936938881874084,
-0.05456335097551346,
-0.1266103982925415
] |
null | null | transformers |
# yuj-v1
The yuj-v1 model is a blend of advanced models strategically crafted to enhance Hindi Language Models (LLMs) effectively and democratically. Its primary goals include catalyzing the development of Hindi and its communities, making significant contributions to linguistic knowledge. The term "yuj," from Sanskrit, signifies fundamental unity, highlighting the integration of sophisticated technologies to improve the language experience for users in the Hindi-speaking community.
Official GGUF version: [shuvom/yuj-v1-GGUF](https://huggingface.co/shuvom/yuj-v1-GGUF)
Below are the model which are leverage to build this yuj-v1:
* [ai4bharat/Airavata](https://huggingface.co/ai4bharat/Airavata)
* [BhabhaAI/Gajendra-v0.1](https://huggingface.co/BhabhaAI/Gajendra-v0.1)
## 🧩 Configuration
```yaml
models:
- model: sarvamai/OpenHathi-7B-Hi-v0.1-Base
# no parameters necessary for base model
- model: ai4bharat/Airavata
parameters:
density: 0.5
weight: 0.5
- model: BhabhaAI/Gajendra-v0.1
parameters:
density: 0.5
weight: 0.3
merge_method: ties
base_model: sarvamai/OpenHathi-7B-Hi-v0.1-Base
parameters:
normalize: true
dtype: float16
```
## 💻 Usage
First, you need to install some of below packages:
1. Bits and bytes
```python
!pip install bitsandbytes
```
2. Accelerate (to install the latest version)
```python
!pip install git+https://github.com/huggingface/accelerate.git
```
3. Usage
```python
# Usage
import torch
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
# load the model in 4-bit quantization
tokenizer = AutoTokenizer.from_pretrained("shuvom/yuj-v1")
model = AutoModelForCausalLM.from_pretrained("shuvom/yuj-v1",torch_dtype=torch.bfloat16,load_in_4bit=True)
prompt = "युज शीर्ष द्विभाषी मॉडल में से एक है"
inputs = tokenizer(prompt, return_tensors="pt")
# Generate
generate_ids = model.generate(inputs.input_ids, max_length=65)
tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
```
4. Output
```python
युज शीर्ष द्विभाषी मॉडल में से एक है। यह एक उत्पादक मॉडल है जो एक साथ एक ट्रांसफॉर्मर और एक आत्म-ध्यान तंत्रिका नेटवर्क को जोड़ता है। यह एक ट्रांसफॉर्मर वास्तुकला का उपयोग करता है जो एक ट्रांसफॉर्मर मॉडल की तुलना में बहुत अधिक जटिल है।
``` | {"license": "apache-2.0", "tags": ["merge", "hindi", "english", "Llama2", "ai4bharat/Airavata", "BhabhaAI/Gajendra-v0.1"]} | text-generation | shuvom/yuj-v1 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"merge",
"hindi",
"english",
"Llama2",
"ai4bharat/Airavata",
"BhabhaAI/Gajendra-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T08:11:57+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #merge #hindi #english #Llama2 #ai4bharat/Airavata #BhabhaAI/Gajendra-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# yuj-v1
The yuj-v1 model is a blend of advanced models strategically crafted to enhance Hindi Language Models (LLMs) effectively and democratically. Its primary goals include catalyzing the development of Hindi and its communities, making significant contributions to linguistic knowledge. The term "yuj," from Sanskrit, signifies fundamental unity, highlighting the integration of sophisticated technologies to improve the language experience for users in the Hindi-speaking community.
Official GGUF version: shuvom/yuj-v1-GGUF
Below are the model which are leverage to build this yuj-v1:
* ai4bharat/Airavata
* BhabhaAI/Gajendra-v0.1
## Configuration
## Usage
First, you need to install some of below packages:
1. Bits and bytes
2. Accelerate (to install the latest version)
3. Usage
4. Output
| [
"# yuj-v1\n\nThe yuj-v1 model is a blend of advanced models strategically crafted to enhance Hindi Language Models (LLMs) effectively and democratically. Its primary goals include catalyzing the development of Hindi and its communities, making significant contributions to linguistic knowledge. The term \"yuj,\" from Sanskrit, signifies fundamental unity, highlighting the integration of sophisticated technologies to improve the language experience for users in the Hindi-speaking community.\n\nOfficial GGUF version: shuvom/yuj-v1-GGUF\n\nBelow are the model which are leverage to build this yuj-v1:\n* ai4bharat/Airavata\n* BhabhaAI/Gajendra-v0.1",
"## Configuration",
"## Usage\n\nFirst, you need to install some of below packages:\n\n1. Bits and bytes\n\n2. Accelerate (to install the latest version)\n\n3. Usage\n\n4. Output"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #merge #hindi #english #Llama2 #ai4bharat/Airavata #BhabhaAI/Gajendra-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# yuj-v1\n\nThe yuj-v1 model is a blend of advanced models strategically crafted to enhance Hindi Language Models (LLMs) effectively and democratically. Its primary goals include catalyzing the development of Hindi and its communities, making significant contributions to linguistic knowledge. The term \"yuj,\" from Sanskrit, signifies fundamental unity, highlighting the integration of sophisticated technologies to improve the language experience for users in the Hindi-speaking community.\n\nOfficial GGUF version: shuvom/yuj-v1-GGUF\n\nBelow are the model which are leverage to build this yuj-v1:\n* ai4bharat/Airavata\n* BhabhaAI/Gajendra-v0.1",
"## Configuration",
"## Usage\n\nFirst, you need to install some of below packages:\n\n1. Bits and bytes\n\n2. Accelerate (to install the latest version)\n\n3. Usage\n\n4. Output"
] | [
88,
161,
4,
38
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #merge #hindi #english #Llama2 #ai4bharat/Airavata #BhabhaAI/Gajendra-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# yuj-v1\n\nThe yuj-v1 model is a blend of advanced models strategically crafted to enhance Hindi Language Models (LLMs) effectively and democratically. Its primary goals include catalyzing the development of Hindi and its communities, making significant contributions to linguistic knowledge. The term \"yuj,\" from Sanskrit, signifies fundamental unity, highlighting the integration of sophisticated technologies to improve the language experience for users in the Hindi-speaking community.\n\nOfficial GGUF version: shuvom/yuj-v1-GGUF\n\nBelow are the model which are leverage to build this yuj-v1:\n* ai4bharat/Airavata\n* BhabhaAI/Gajendra-v0.1## Configuration## Usage\n\nFirst, you need to install some of below packages:\n\n1. Bits and bytes\n\n2. Accelerate (to install the latest version)\n\n3. Usage\n\n4. Output"
] | [
-0.05991892144083977,
-0.013398412615060806,
-0.0013414706336334348,
0.020102599635720253,
0.05986668914556503,
0.02006431855261326,
0.02989206649363041,
0.11537489295005798,
-0.04559772461652756,
0.033357422798871994,
0.09335760027170181,
0.03814932331442833,
0.07868017256259918,
0.21958912909030914,
0.024861659854650497,
-0.16082246601581573,
0.040586866438388824,
-0.029023000970482826,
-0.08284232765436172,
0.11853723227977753,
0.11632195115089417,
-0.030726449564099312,
0.0312567800283432,
0.0037807892076671124,
-0.13886232674121857,
-0.046683646738529205,
-0.03674580156803131,
-0.0293875802308321,
0.04296555742621422,
0.027549609541893005,
0.043341461569070816,
0.028498586267232895,
-0.00021594441204797477,
-0.19905206561088562,
0.02502557262778282,
-0.009648817591369152,
-0.03272077813744545,
0.005967752076685429,
-0.0007182893459685147,
0.06842908263206482,
0.19877664744853973,
-0.07229312509298325,
-0.06086399406194687,
0.07598876953125,
-0.07585897296667099,
-0.0984417274594307,
-0.03264862671494484,
0.03844115883111954,
0.1228330209851265,
0.05525726452469826,
-0.05022238567471504,
0.09806502610445023,
-0.05374511703848839,
0.041617490351200104,
-0.07316233217716217,
-0.12391875684261322,
-0.032990168780088425,
0.08433603495359421,
0.06954129040241241,
0.06171542406082153,
-0.004502270836383104,
0.028087155893445015,
0.062237683683633804,
-0.005836824886500835,
0.08402987569570541,
-0.027410125359892845,
0.17433029413223267,
-0.013075954280793667,
-0.11203297972679138,
-0.047940004616975784,
0.13150028884410858,
-0.030201872810721397,
-0.018721723929047585,
-0.1372053027153015,
-0.04519227519631386,
0.07216432690620422,
-0.02311858721077442,
-0.07423361390829086,
-0.016483211889863014,
0.09611796587705612,
0.142002671957016,
-0.026371721178293228,
-0.09418552368879318,
0.05039263516664505,
0.10557054728269577,
0.09723983705043793,
0.07915326952934265,
0.056415751576423645,
-0.018704542890191078,
-0.0041218968108296394,
-0.0947558581829071,
-0.126039519906044,
-0.023190079256892204,
-0.09403537213802338,
0.02806933969259262,
0.04545428231358528,
0.023227600380778313,
0.002479556482285261,
0.07931329309940338,
0.03109133057296276,
-0.1386515498161316,
-0.010417962446808815,
0.04147924482822418,
-0.006200646515935659,
0.06854040175676346,
0.005581831093877554,
-0.03210820257663727,
-0.0210897084325552,
0.09775634109973907,
0.038951024413108826,
0.07649102807044983,
-0.009740617126226425,
-0.05192800238728523,
0.036419183015823364,
0.004077829886227846,
0.06859573721885681,
0.04115906357765198,
0.008469048887491226,
-0.04716571047902107,
-0.06349936127662659,
0.13785484433174133,
-0.1000087708234787,
-0.02564927563071251,
-0.061437997967004776,
0.03475099056959152,
0.016453411430120468,
-0.010309383273124695,
-0.017861582338809967,
-0.0688011646270752,
-0.08286205679178238,
-0.05713373050093651,
0.011686081066727638,
-0.07969653606414795,
-0.05885796621441841,
-0.00437812739983201,
-0.13429602980613708,
0.0035133317578583956,
-0.08188747614622116,
-0.14387363195419312,
0.013049708679318428,
0.07090894877910614,
-0.09360053390264511,
-0.07061069458723068,
-0.052736442536115646,
-0.08108998090028763,
0.0565376915037632,
-0.03133212402462959,
-0.03323787450790405,
-0.03827953711152077,
0.04422919079661369,
0.07164705544710159,
0.05160186439752579,
0.03496968373656273,
0.021565280854701996,
-0.02778414450585842,
0.0786544606089592,
0.0033335068728774786,
0.05631275847554207,
-0.09048279374837875,
0.01818622648715973,
-0.09451603144407272,
-0.010727477259933949,
0.12546901404857635,
0.06920218467712402,
-0.021014582365751266,
0.17140701413154602,
-0.12307561933994293,
-0.045974068343639374,
0.1430405080318451,
-0.21153756976127625,
-0.12333443015813828,
0.07040654122829437,
0.040101293474435806,
0.158577561378479,
0.08447270095348358,
0.035282786935567856,
0.10545294731855392,
-0.085829958319664,
-0.13165947794914246,
0.015867121517658234,
-0.004673273768275976,
0.004999450407922268,
0.06279589235782623,
0.09326110780239105,
0.07274099439382553,
0.03254448249936104,
-0.0509517565369606,
0.09514576941728592,
-0.04015873372554779,
-0.06744670122861862,
-0.04961115121841431,
-0.06081525236368179,
0.03958924487233162,
0.0060090310871601105,
0.04429013654589653,
0.0035438521299511194,
-0.12593385577201843,
-0.06977437436580658,
0.11633020639419556,
-0.03802526369690895,
0.04454405978322029,
-0.12715841829776764,
0.08477532863616943,
-0.09436163306236267,
0.0554451085627079,
-0.023677879944443703,
0.08795024454593658,
0.07679862529039383,
-0.03835705667734146,
-0.022012759000062943,
-0.058683112263679504,
0.038842108100652695,
0.031168257817626,
-0.05011209845542908,
-0.03183334320783615,
0.02596881240606308,
-0.017552347853779793,
0.017483830451965332,
-0.15867497026920319,
-0.003424766007810831,
-0.06588687747716904,
0.050168536603450775,
-0.08800265938043594,
0.03352734446525574,
0.1404096484184265,
0.05451655015349388,
-0.03298502787947655,
-0.006276983302086592,
0.0608956441283226,
0.0080890953540802,
-0.04239911958575249,
-0.03236117586493492,
0.045621614903211594,
-0.04294614866375923,
-0.09418909251689911,
0.13510660827159882,
-0.032586656510829926,
-0.013389606960117817,
0.09292971342802048,
-0.06268811970949173,
0.06599107384681702,
0.05372704565525055,
-0.021298963576555252,
-0.06605914980173111,
0.06306954473257065,
-0.04929935559630394,
-0.014168131165206432,
-0.013261101208627224,
0.08578062057495117,
-0.053548555821180344,
-0.06739868968725204,
0.033497266471385956,
-0.045902594923973083,
-0.030921386554837227,
-0.011791977100074291,
0.00960297416895628,
-0.015030751936137676,
0.15639089047908783,
0.11316840350627899,
0.028688661754131317,
0.19813676178455353,
-0.0640946701169014,
-0.05427362397313118,
-0.02382281981408596,
0.08057484775781631,
-0.02301814593374729,
0.06436273455619812,
-0.17750926315784454,
-0.05310380831360817,
0.02142108976840973,
0.007653460372239351,
0.03795723244547844,
-0.0691729187965393,
-0.010631747543811798,
-0.026539117097854614,
-0.060457076877355576,
0.05849403515458107,
0.02480596862733364,
-0.10080838203430176,
0.03893367201089859,
0.010491825640201569,
0.12075518071651459,
0.050583623349666595,
0.003826113184913993,
-0.09865181893110275,
0.11752070486545563,
-0.15717938542366028,
-0.30908605456352234,
-0.13855011761188507,
-0.15151573717594147,
-0.013300731778144836,
0.01838815025985241,
0.05747843161225319,
-0.21473261713981628,
-0.10001722723245621,
0.015007168054580688,
0.1445334404706955,
-0.05920146778225899,
0.01776670478284359,
-0.007490594871342182,
0.02590227872133255,
0.011566679924726486,
-0.11966928839683533,
0.014059578999876976,
0.05248066410422325,
-0.06824176013469696,
0.12058219313621521,
0.04986384138464928,
0.056646816432476044,
0.027183780446648598,
0.0359688438475132,
-0.017551111057400703,
-0.024214237928390503,
0.23071445524692535,
-0.08522489666938782,
0.1221725270152092,
0.2060239613056183,
-0.02649567276239395,
0.045832160860300064,
0.18048088252544403,
0.009081680327653885,
-0.06404915452003479,
-0.017491916194558144,
-0.12123221158981323,
-0.05732977017760277,
-0.1966133713722229,
-0.11253073066473007,
-0.06785822659730911,
-0.047256436198949814,
-0.002492576837539673,
0.053667716681957245,
0.0068313563242554665,
0.10878775268793106,
-0.0858733281493187,
0.1014876738190651,
0.057029224932193756,
0.06991035491228104,
0.1456589251756668,
-0.025637982413172722,
0.10056035965681076,
0.0003793841751758009,
0.03764438256621361,
0.10041198879480362,
0.022662553936243057,
0.2865123152732849,
0.005670467857271433,
0.15134185552597046,
0.15273195505142212,
0.14876161515712738,
0.08270137012004852,
-0.008132236078381538,
-0.04572579264640808,
0.017946740612387657,
-0.010445752181112766,
-0.06273549050092697,
-0.05049876496195793,
0.14418631792068481,
-0.05373133346438408,
-0.0114962263032794,
0.09225527942180634,
0.024724701419472694,
0.045829348266124725,
0.09487908333539963,
0.018159354105591774,
-0.08032964915037155,
-0.10712164640426636,
0.09667669236660004,
-0.08367890864610672,
-0.03143320977687836,
-0.010868378914892673,
0.07851644605398178,
-0.0021590865217149258,
0.0641632005572319,
-0.04092010483145714,
0.04588264599442482,
-0.04469660297036171,
0.00833770353347063,
-0.0824134424328804,
0.005979036446660757,
0.05707408860325813,
0.09668685495853424,
-0.3028728663921356,
0.16860945522785187,
0.04968588054180145,
0.05571317672729492,
-0.025347989052534103,
0.11156304180622101,
-0.022042032331228256,
0.004249821417033672,
0.08480758219957352,
0.0019852984696626663,
0.0311114564538002,
-0.06041831895709038,
-0.08905281126499176,
0.03401808813214302,
0.04578496888279915,
0.023053230717778206,
0.0737156867980957,
-0.021862948313355446,
0.027641376480460167,
-0.026434432715177536,
0.031157104298472404,
-0.1688227355480194,
-0.11040127277374268,
0.09827816486358643,
-0.05123164504766464,
0.04708340764045715,
-0.07108037173748016,
-0.053634777665138245,
-0.011603659950196743,
0.15368279814720154,
-0.17465801537036896,
-0.1594424843788147,
-0.04241138696670532,
0.06251464039087296,
0.05876346677541733,
-0.08315256237983704,
0.06280873715877533,
0.011881394311785698,
0.005097603891044855,
-0.05106071010231972,
-0.0312819741666317,
-0.006308108568191528,
-0.0936962440609932,
-0.12077580392360687,
-0.01104759145528078,
0.11074646562337875,
0.08562959730625153,
-0.017432456836104393,
-0.0071928189136087894,
0.0459686815738678,
-0.04238211363554001,
-0.09393635392189026,
-0.0054890853352844715,
0.19199493527412415,
-0.023574771359562874,
-0.005890995729714632,
-0.051975466310977936,
-0.1389482468366623,
-0.03485061973333359,
-0.15989968180656433,
0.06982213258743286,
0.33988237380981445,
-0.005139425862580538,
0.23297828435897827,
0.19545024633407593,
-0.08959991484880447,
-0.1926954835653305,
-0.16262125968933105,
0.017217839136719704,
0.0636601597070694,
-0.0040716384537518024,
-0.15349078178405762,
0.0332706943154335,
-0.005408541299402714,
-0.04494982212781906,
-0.009416408836841583,
-0.07336869835853577,
-0.07262799888849258,
-0.01874685473740101,
0.13142366707324982,
0.1745375245809555,
-0.147142693400383,
-0.061564989387989044,
-0.006351469550281763,
0.055437106639146805,
0.0221769530326128,
0.01654118113219738,
0.0013272024225443602,
-0.05486780405044556,
-0.035253092646598816,
-0.0038935684133321047,
-0.035862527787685394,
0.10365213453769684,
-0.06357219070196152,
-0.0690508484840393,
-0.07801345735788345,
0.08637283742427826,
0.19129078090190887,
-0.005742068402469158,
0.05092630907893181,
-0.07373793423175812,
-0.02144731394946575,
-0.13349325954914093,
-0.023063089698553085,
-0.027272703126072884,
0.03853592276573181,
-0.044512730091810226,
-0.07915662974119186,
-0.042643092572689056,
0.047688648104667664,
0.05607195198535919,
0.019053637981414795,
0.025846149772405624,
-0.004213130567222834,
0.01948830857872963,
0.19560329616069794,
0.11183610558509827,
-0.12487392127513885,
0.11432571709156036,
-0.028160996735095978,
0.006989135406911373,
0.029097450897097588,
-0.2770021855831146,
-0.017117226496338844,
0.004958086647093296,
-0.015707822516560555,
0.08170699328184128,
-0.01377470139414072,
-0.06864683330059052,
0.0804109126329422,
0.13285207748413086,
0.019704418256878853,
-0.12176111340522766,
-0.059317342936992645,
-0.09921451658010483,
-0.03196790814399719,
-0.04556502029299736,
0.1300579458475113,
-0.055816926062107086,
-0.04221624135971069,
-0.04837840422987938,
0.0731031745672226,
-0.043803077191114426,
0.09174591302871704,
0.01899142377078533,
0.06337562203407288,
-0.07784144580364227,
-0.017461061477661133,
0.05533717945218086,
-0.04774731397628784,
0.024746887385845184,
0.1337721049785614,
-0.12629379332065582,
-0.08824144303798676,
-0.055234394967556,
-0.12040448188781738,
0.07797389477491379,
-0.007798139471560717,
0.004558140411973,
-0.03907518461346626,
0.044936902821063995,
-0.036807458847761154,
0.029949618503451347,
0.017560143023729324,
-0.023264212533831596,
0.04189678654074669,
-0.05614100396633148,
0.1150704100728035,
0.035550907254219055,
-0.005986477714031935,
-0.1735485941171646,
-0.10316688567399979,
0.0767611488699913,
0.09002932906150818,
-0.06652501970529556,
-0.016777219250798225,
-0.14646916091442108,
-0.007596517447382212,
-0.11398141831159592,
0.09307268261909485,
-0.14804184436798096,
-0.03286828100681305,
-0.030731813982129097,
-0.060506537556648254,
-0.029702497646212578,
0.06749303638935089,
-0.026185765862464905,
-0.00414457218721509,
-0.05559796094894409,
0.1053885668516159,
-0.010103404521942139,
-0.00199010968208313,
0.06857037544250488,
-0.06503186374902725,
0.07722801715135574,
-0.038616977632045746,
-0.06234540790319443,
0.08956190943717957,
0.006511083338409662,
0.03330365940928459,
0.030374664813280106,
0.008773710578680038,
0.004660970997065306,
-0.12487790733575821,
0.00042433428461663425,
0.060932159423828125,
-0.03118492290377617,
0.027417199686169624,
0.0938316360116005,
-0.018397672101855278,
0.038570091128349304,
-0.0682150200009346,
-0.006184098776429892,
-0.015860334038734436,
0.0728486105799675,
0.11666249483823776,
0.0097196651622653,
0.12511079013347626,
-0.05421704798936844,
0.022940780967473984,
-0.137245774269104,
0.012819001451134682,
-0.016406605020165443,
-0.008379998616874218,
-0.14115241169929504,
-0.10009787231683731,
-0.009110485203564167,
-0.029786957427859306,
0.13649672269821167,
-0.0449586845934391,
0.004387179855257273,
0.03664257377386093,
-0.037777550518512726,
-0.06320973485708237,
-0.008421008475124836,
0.202156662940979,
0.09982433170080185,
0.08004187047481537,
-0.053398966789245605,
-0.00398925831541419,
-0.00213532499037683,
-0.18071241676807404,
-0.1767151951789856,
0.025787141174077988,
-0.015178261324763298,
0.04641468822956085,
-0.005153636913746595,
-0.03296357020735741,
-0.1552957147359848,
0.12754938006401062,
-0.10872084647417068,
0.032683078199625015,
-0.04468308389186859,
0.17202939093112946,
0.19681252539157867,
-0.1183558776974678,
0.04835270345211029,
0.025534439831972122,
-0.0464896559715271,
-0.040641847997903824,
-0.20538195967674255,
-0.0868971049785614,
-0.20946234464645386,
0.06391500681638718,
-0.11675690114498138,
-0.006094298325479031,
0.11758866906166077,
0.020454691722989082,
0.012227436527609825,
0.04450353607535362,
0.16410914063453674,
-0.0747775286436081,
-0.010641003958880901,
-0.01139360573142767,
-0.012130049988627434,
0.05106547102332115,
0.015936026349663734,
-0.049291618168354034,
0.014291182160377502,
0.04458065330982208,
0.03825505077838898,
-0.020153839141130447,
0.00314629590138793,
-0.1257895827293396,
-0.08305985480546951,
0.014174784533679485,
0.0985482856631279,
0.01642202027142048,
0.15697290003299713,
-0.025880437344312668,
-0.08830717951059341,
0.023663120344281197,
0.14311102032661438,
0.01794147863984108,
-0.1399071216583252,
-0.10875391215085983,
0.14032450318336487,
-0.008434311486780643,
-0.023837465792894363,
0.040554892271757126,
-0.018062615767121315,
-0.021400364115834236,
0.19314360618591309,
0.18316397070884705,
0.06942880153656006,
-0.00911413598805666,
-0.03944668173789978,
0.007181502413004637,
-0.07812248170375824,
0.16537000238895416,
0.09988465160131454,
0.15990372002124786,
-0.08372513204813004,
0.03438650816679001,
-0.08247075229883194,
-0.06896492093801498,
-0.10830067843198776,
0.019092928618192673,
-0.05063658207654953,
-0.0402861051261425,
-0.018760832026600838,
0.08969588577747345,
-0.01090081874281168,
0.0021421804558485746,
-0.04351997748017311,
-0.13674965500831604,
-0.05166833475232124,
-0.042040906846523285,
0.028338229283690453,
-0.01108742505311966,
-0.013828004710376263,
0.03948652371764183,
-0.03164329752326012,
0.08258563280105591,
0.015700628980994225,
-0.09969469904899597,
0.02996792457997799,
0.16161353886127472,
-0.15375873446464539,
0.031282372772693634,
-0.011424881406128407,
0.09429413825273514,
0.06630754470825195,
0.0236747357994318,
-0.09525120258331299,
0.044168680906295776,
0.02143115922808647,
-0.0360061451792717,
-0.020002266392111778,
0.19552752375602722,
-0.032694119960069656,
0.0232971403747797,
0.0726395919919014,
-0.08458627760410309,
-0.027040645480155945,
-0.01467036735266447,
-0.0038536032661795616,
-0.02529050037264824,
0.07581781595945358,
-0.10284928232431412,
0.13109023869037628,
0.05424732714891434,
-0.045576076954603195,
-0.029946403577923775,
-0.10395527631044388,
0.08887878805398941,
-0.049432627856731415,
0.17974726855754852,
-0.01963658817112446,
-0.23455841839313507,
-0.06158103048801422,
0.11525316536426544,
0.07404070347547531,
-0.10180029273033142,
-0.04491814970970154,
-0.055147454142570496,
0.053014401346445084,
-0.13082247972488403,
0.15650568902492523,
0.10969802737236023,
0.007599577773362398,
-0.0373222641646862,
-0.1448921114206314,
-0.009625163860619068,
0.032859738916158676,
-0.10230397433042526,
-0.044622473418712616
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# whisper-tiny-ft-verbatim-cy
This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7686
- Wer: 53.9846
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 4000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:-------:|
| 0.9656 | 1.41 | 1000 | 0.9762 | 67.1825 |
| 0.7198 | 2.83 | 2000 | 0.8223 | 59.9990 |
| 0.5869 | 4.24 | 3000 | 0.7813 | 54.7898 |
| 0.5413 | 5.66 | 4000 | 0.7686 | 53.9846 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "openai/whisper-tiny", "model-index": [{"name": "whisper-tiny-ft-verbatim-cy", "results": []}]} | automatic-speech-recognition | DewiBrynJones/whisper-tiny-ft-verbatim-cy | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:openai/whisper-tiny",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-06T08:14:16+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-openai/whisper-tiny #license-apache-2.0 #endpoints_compatible #region-us
| whisper-tiny-ft-verbatim-cy
===========================
This model is a fine-tuned version of openai/whisper-tiny on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7686
* Wer: 53.9846
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 8
* total\_train\_batch\_size: 32
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* training\_steps: 4000
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.2.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-openai/whisper-tiny #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
68,
143,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-openai/whisper-tiny #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.13621056079864502,
0.12308640778064728,
-0.0010191290639340878,
0.07383094727993011,
0.11999335885047913,
0.006068327464163303,
0.1332564800977707,
0.13278315961360931,
-0.10892076790332794,
0.0736992210149765,
0.09363288432359695,
0.07405585050582886,
0.03676488250494003,
0.14220014214515686,
-0.028328103944659233,
-0.3041890859603882,
0.006419685203582048,
0.010933640412986279,
-0.1288028061389923,
0.12801550328731537,
0.09646142274141312,
-0.12642189860343933,
0.03737591579556465,
0.026124805212020874,
-0.15569348633289337,
-0.0199345164000988,
-0.0023839978966861963,
-0.09620264172554016,
0.12743701040744781,
0.012078220956027508,
0.09189800918102264,
0.03351560980081558,
0.10510635375976562,
-0.1863100528717041,
0.0071166083216667175,
0.0585462860763073,
0.04091288521885872,
0.09463286399841309,
0.06554010510444641,
0.004474124871194363,
0.08878681063652039,
-0.04652651399374008,
0.08637333661317825,
0.048006847500801086,
-0.09477119147777557,
-0.3300764858722687,
-0.09350567311048508,
0.08681253343820572,
0.11805450916290283,
0.08149585127830505,
-0.0160340778529644,
0.12991049885749817,
-0.06527054309844971,
0.08235844224691391,
0.2516932487487793,
-0.28114011883735657,
-0.08221140503883362,
-0.030504044145345688,
0.054762933403253555,
0.016959285363554955,
-0.11951996386051178,
-0.014028883539140224,
0.05597937852144241,
0.03248630091547966,
0.10170560330152512,
0.008655289188027382,
-0.0013880531769245863,
-0.009865351021289825,
-0.13940227031707764,
-0.052768245339393616,
0.11945780366659164,
0.05414256826043129,
-0.055454008281230927,
-0.10770418494939804,
-0.04851971194148064,
-0.1865122765302658,
-0.048182617872953415,
-0.006607586517930031,
0.03234073892235756,
-0.07585299760103226,
-0.14501908421516418,
0.018121611326932907,
-0.08689919114112854,
-0.10985691100358963,
0.020633161067962646,
0.21591916680335999,
0.04981570690870285,
-0.0017913804622367024,
-0.017667775973677635,
0.11500908434391022,
0.048070281744003296,
-0.17306455969810486,
0.0019028817769140005,
0.03212805837392807,
-0.0449579618871212,
-0.007688626181334257,
-0.04883097484707832,
-0.020864788442850113,
0.012348102405667305,
0.17717772722244263,
-0.07237683236598969,
0.04051290079951286,
0.026729684323072433,
0.03053714521229267,
-0.09945792704820633,
0.2005079686641693,
-0.07399120926856995,
-0.032049886882305145,
-0.013439259491860867,
0.13086234033107758,
0.05162305012345314,
-0.01518266461789608,
-0.07998824119567871,
0.02796083129942417,
0.09871228784322739,
0.04911600053310394,
-0.012868293561041355,
0.03528422489762306,
-0.030276279896497726,
-0.020220208913087845,
0.039441462606191635,
-0.09825509786605835,
0.021194668486714363,
0.011335700750350952,
-0.09278415888547897,
-0.0390763096511364,
0.007713764440268278,
0.01645267941057682,
0.002081101294606924,
0.13833558559417725,
-0.08719552308320999,
-0.01543380692601204,
-0.08647765964269638,
-0.08495502918958664,
0.016892602667212486,
-0.023624606430530548,
-0.0008343344670720398,
-0.08473587781190872,
-0.11597950756549835,
-0.02106558531522751,
0.03454214707016945,
-0.03347054496407509,
-0.07348936051130295,
-0.03464273363351822,
-0.1116122305393219,
0.03512237221002579,
-0.013980173505842686,
0.11056531220674515,
-0.043882161378860474,
0.12242833524942398,
0.07126748561859131,
0.058151137083768845,
0.031352005898952484,
0.0378434881567955,
-0.08557241410017014,
0.039330463856458664,
-0.1986258327960968,
0.0406634695827961,
-0.07555270940065384,
0.07505030184984207,
-0.10022484511137009,
-0.13256599009037018,
0.03468246012926102,
-0.008948971517384052,
0.09543298929929733,
0.12583012878894806,
-0.15313516557216644,
-0.10731427371501923,
0.18446415662765503,
-0.11627252399921417,
-0.13991253077983856,
0.1338636428117752,
-0.0028976884204894304,
-0.016717107966542244,
0.03331249952316284,
0.16532009840011597,
0.07892947643995285,
-0.09537186473608017,
-0.024019738659262657,
-0.042915310710668564,
0.11110378801822662,
-0.0035669647622853518,
0.10013725608587265,
-0.02834591083228588,
0.04073804244399071,
0.02260737121105194,
-0.04819858819246292,
0.03845541924238205,
-0.10522440075874329,
-0.09258246421813965,
-0.03497782349586487,
-0.10158446431159973,
0.06230342015624046,
0.07215838879346848,
0.061382003128528595,
-0.11332394182682037,
-0.12745322287082672,
0.033378008753061295,
0.130001038312912,
-0.07410112023353577,
0.02120630256831646,
-0.09479771554470062,
0.08260861039161682,
-0.03160678967833519,
-0.014541469514369965,
-0.16732802987098694,
-0.009080324321985245,
0.03617598116397858,
-0.05041569471359253,
-0.0010080840438604355,
-0.06539595127105713,
0.08283466100692749,
0.078203946352005,
-0.0760059729218483,
-0.07834390550851822,
-0.08319555222988129,
-0.0153697794303298,
-0.0766870304942131,
-0.22780083119869232,
-0.09752421826124191,
-0.02567426487803459,
0.15675760805606842,
-0.1955004334449768,
0.03707297891378403,
0.011206614784896374,
0.14156608283519745,
0.04049690067768097,
-0.03599758446216583,
-0.010603123344480991,
0.07288811355829239,
-0.017694758251309395,
-0.060830745846033096,
0.02899128384888172,
0.014148597605526447,
-0.11043106019496918,
-0.01389783900231123,
-0.12862516939640045,
0.15469543635845184,
0.12204582989215851,
0.011776819825172424,
-0.08844760805368423,
0.006660010199993849,
-0.09871510416269302,
-0.04328417032957077,
-0.02598140947520733,
-0.01288729440420866,
0.12685437500476837,
0.02686995454132557,
0.1352800726890564,
-0.089844711124897,
-0.06796985119581223,
0.03856012597680092,
-0.021409060806035995,
-0.003656940534710884,
0.10636994242668152,
0.022764459252357483,
-0.027108021080493927,
0.120271235704422,
0.10228756070137024,
-0.09780120104551315,
0.13699866831302643,
-0.08233048021793365,
-0.08612639456987381,
-0.02103067934513092,
0.005667049437761307,
0.04337294399738312,
0.14442020654678345,
-0.0902925431728363,
-0.0174952894449234,
0.019106538966298103,
-0.010542538948357105,
0.017264194786548615,
-0.2125503420829773,
-0.015406347811222076,
0.025332747027277946,
-0.04674898833036423,
-0.02641904726624489,
-0.01081561017781496,
0.002304854104295373,
0.10242942720651627,
0.010280468501150608,
-0.05111363157629967,
0.006934516131877899,
-0.0064778560772538185,
-0.06335088610649109,
0.20417219400405884,
-0.07774090021848679,
-0.14120183885097504,
-0.15146814286708832,
0.018421294167637825,
-0.042422208935022354,
-0.0024780048988759518,
0.04458596929907799,
-0.11058292537927628,
-0.01294918917119503,
-0.056690022349357605,
0.0316988006234169,
-0.005696469452232122,
0.04638686403632164,
0.029616670683026314,
0.03116745688021183,
0.09627756476402283,
-0.11580602079629898,
0.028350651264190674,
-0.04133334383368492,
-0.05350249260663986,
0.017807453870773315,
0.05936488136649132,
0.10235635936260223,
0.1730482280254364,
0.0035602988209575415,
0.01250386517494917,
-0.030978385359048843,
0.1637604981660843,
-0.11215642839670181,
-0.042417507618665695,
0.15017981827259064,
0.00005088376929052174,
0.05127294361591339,
0.13505497574806213,
0.06416266411542892,
-0.08256293833255768,
0.009197883307933807,
0.023735778406262398,
-0.02192429080605507,
-0.21619023382663727,
-0.03499075397849083,
-0.028703169897198677,
0.028044160455465317,
0.09736949950456619,
0.03359614685177803,
0.03483157604932785,
0.045335207134485245,
-0.03217777982354164,
0.00013004642096348107,
-0.026757176965475082,
0.07812588661909103,
0.05535273253917694,
0.042670201510190964,
0.12948478758335114,
-0.02758748270571232,
-0.054917868226766586,
0.01592368073761463,
-0.025599217042326927,
0.1927938163280487,
-0.0423513799905777,
0.13387757539749146,
0.042872775346040726,
0.1516958475112915,
0.018484849482774734,
0.07975566387176514,
0.013422983698546886,
-0.03772109001874924,
0.02538190968334675,
-0.06382511556148529,
-0.03882383555173874,
0.03310453146696091,
0.010604211129248142,
0.1029682606458664,
-0.1390507072210312,
0.00307364109903574,
0.05206160247325897,
0.3518446981906891,
0.07149916887283325,
-0.31084808707237244,
-0.14059437811374664,
-0.0007493052980862558,
-0.07699736952781677,
-0.023943059146404266,
0.03015109710395336,
0.13964447379112244,
-0.09540048986673355,
0.05627627670764923,
-0.07880900800228119,
0.0762789323925972,
-0.03906766325235367,
0.022476904094219208,
0.08232531696557999,
0.08865737169981003,
-0.0026578165125101805,
0.04832261800765991,
-0.2400757223367691,
0.30755528807640076,
-0.005839981138706207,
0.09712006151676178,
-0.041227128356695175,
0.0325959138572216,
0.025828780606389046,
0.012107371352612972,
0.07370225340127945,
-0.026632236316800117,
-0.054221153259277344,
-0.18540741503238678,
-0.08071563392877579,
0.018031248822808266,
0.13722801208496094,
-0.06736595183610916,
0.12585245072841644,
-0.02633794955909252,
-0.029028255492448807,
0.048902902752161026,
-0.06593510508537292,
-0.05800614506006241,
-0.07103220373392105,
0.023513313382864,
0.003128496464341879,
0.033872611820697784,
-0.12921889126300812,
-0.12476766854524612,
-0.052745841443538666,
0.13977670669555664,
-0.08339928090572357,
-0.048023711889982224,
-0.1243181899189949,
0.07167290896177292,
0.16487184166908264,
-0.0703386589884758,
0.0582202672958374,
0.013004747219383717,
0.13503466546535492,
0.0058110603131353855,
-0.02054525353014469,
0.09166724234819412,
-0.08879637718200684,
-0.26695069670677185,
-0.04963972419500351,
0.18234136700630188,
0.006408292334526777,
0.0664517804980278,
-0.02236476168036461,
0.033265672624111176,
-0.0038914470933377743,
-0.08053449541330338,
0.0436432883143425,
-0.012259279377758503,
0.02841208316385746,
0.007611429784446955,
-0.012324435636401176,
-0.0010882256319746375,
-0.06594711542129517,
-0.03712284564971924,
0.12417453527450562,
0.2987803518772125,
-0.09523279219865799,
0.006108872592449188,
0.0991329699754715,
-0.018491897732019424,
-0.1665424257516861,
0.012641362845897675,
0.11193069815635681,
0.01853560283780098,
-0.011049405671656132,
-0.1759742647409439,
0.07983607053756714,
0.08755503594875336,
-0.05739418789744377,
0.08933739364147186,
-0.31361475586891174,
-0.1445634812116623,
0.10107993334531784,
0.1080315038561821,
-0.006482536904513836,
-0.16973325610160828,
-0.06478404253721237,
-0.01751735806465149,
-0.08526802062988281,
0.06286703050136566,
-0.06577438116073608,
0.1124669536948204,
-0.012351091019809246,
0.050366874784231186,
0.024494023993611336,
-0.06776367872953415,
0.1285332590341568,
-0.03095719777047634,
0.07386243343353271,
-0.01808890327811241,
-0.004944432992488146,
0.0682871863245964,
-0.04944641515612602,
0.009474622085690498,
-0.06250214576721191,
0.047247856855392456,
-0.06779391318559647,
-0.01654166169464588,
-0.10675714164972305,
0.03198259323835373,
-0.04977655038237572,
-0.06327582150697708,
-0.008488301187753677,
0.04717617109417915,
0.018344905227422714,
-0.020190328359603882,
0.13811184465885162,
-0.0024739555083215237,
0.2137630581855774,
0.12076349556446075,
0.08659151941537857,
-0.02469480223953724,
-0.0602940134704113,
0.004351360257714987,
-0.029234450310468674,
0.07441405206918716,
-0.1380317062139511,
0.015881413593888283,
0.12333464622497559,
0.08265487849712372,
0.1023174300789833,
0.06724931299686432,
-0.07170028984546661,
0.01947692409157753,
0.08603265881538391,
-0.1277482807636261,
-0.1265818029642105,
-0.022663326933979988,
0.03951262682676315,
-0.1651885062456131,
0.08795836567878723,
0.09540806710720062,
-0.07661629468202591,
-0.019686337560415268,
0.003545002778992057,
0.0065987855195999146,
-0.03538352623581886,
0.23720191419124603,
0.06767157465219498,
0.09885862469673157,
-0.10651928931474686,
0.09031455963850021,
0.033317096531391144,
-0.1313251405954361,
0.018735062330961227,
0.09227922558784485,
-0.05757548660039902,
-0.011450418271124363,
0.0073831574991345406,
0.03577348217368126,
0.002719932235777378,
-0.05416521430015564,
-0.16937577724456787,
-0.12930943071842194,
0.05635247007012367,
0.15444520115852356,
0.06071622297167778,
0.030293092131614685,
-0.02351979725062847,
0.05671478435397148,
-0.14396125078201294,
0.1195758804678917,
0.062162626534700394,
0.08283868432044983,
-0.15473498404026031,
0.18961383402347565,
0.008200256153941154,
0.03149406611919403,
-0.011813205666840076,
0.019727328792214394,
-0.11354196071624756,
0.01832946389913559,
-0.10816311091184616,
-0.04474984109401703,
-0.04774994030594826,
-0.008585064671933651,
-0.007064737845212221,
-0.0441739447414875,
-0.0640195906162262,
0.03090800903737545,
-0.11817680299282074,
-0.04179324209690094,
0.013714699074625969,
0.023555580526590347,
-0.1198507696390152,
-0.0002258748863823712,
0.043128978461027145,
-0.11388297379016876,
0.10042253881692886,
0.06084232032299042,
0.03271547704935074,
0.04904060438275337,
-0.06652995198965073,
-0.008290090598165989,
0.04129244014620781,
-0.015292760916054249,
0.03735761344432831,
-0.11637608706951141,
-0.018754251301288605,
-0.03625374287366867,
0.032460104674100876,
0.002350590890273452,
0.06210915371775627,
-0.14608922600746155,
-0.004097436089068651,
0.012881563045084476,
-0.045075368136167526,
-0.06740116328001022,
0.020420921966433525,
0.07556503266096115,
0.009397742338478565,
0.15088561177253723,
-0.08594658970832825,
0.039496783167123795,
-0.22628214955329895,
0.0007091375300660729,
-0.035092078149318695,
-0.09954672306776047,
-0.11170665919780731,
0.010766102001070976,
0.079366035759449,
-0.06072615832090378,
0.0833403617143631,
-0.061228908598423004,
0.09898297488689423,
0.04429415613412857,
-0.06584750860929489,
0.02640909142792225,
0.05591046065092087,
0.21566610038280487,
0.03673810139298439,
-0.029630931094288826,
0.0622650571167469,
0.015338777564466,
0.051679715514183044,
0.08350003510713577,
0.1501314789056778,
0.13861557841300964,
0.03457232192158699,
0.09109491854906082,
0.07918019592761993,
-0.0733725056052208,
-0.18466144800186157,
0.04006737470626831,
-0.02902647852897644,
0.10882070660591125,
-0.02024816907942295,
0.18559673428535461,
0.13693679869174957,
-0.15765300393104553,
0.044236183166503906,
-0.04387999325990677,
-0.0657404214143753,
-0.09736104309558868,
-0.016059594228863716,
-0.0674741119146347,
-0.17368102073669434,
0.018187373876571655,
-0.11999593675136566,
0.017692269757390022,
0.07258597016334534,
0.01578211970627308,
0.010885391384363174,
0.1655605435371399,
0.033503465354442596,
0.03664320334792137,
0.09638126194477081,
0.02501453272998333,
-0.024730734527111053,
-0.027229532599449158,
-0.09247103333473206,
0.026286356151103973,
-0.013262470252811909,
0.044625721871852875,
-0.06330430507659912,
-0.1360594928264618,
0.06075052171945572,
0.02295473963022232,
-0.11079816520214081,
0.03837820887565613,
-0.002850443357601762,
0.0978919044137001,
0.029599107801914215,
0.014895625412464142,
0.004865796770900488,
-0.026744725182652473,
0.27461135387420654,
-0.13434487581253052,
-0.08729546517133713,
-0.14443357288837433,
0.29766032099723816,
-0.00122106506023556,
-0.028810344636440277,
0.027034340426325798,
-0.09849037230014801,
-0.042444583028554916,
0.1638517677783966,
0.17240974307060242,
-0.010716243647038937,
-0.012958981096744537,
0.016401782631874084,
-0.019615493714809418,
-0.09077763557434082,
0.07432674616575241,
0.11954616755247116,
0.09712126106023788,
-0.06965260952711105,
-0.03429245576262474,
-0.02953464351594448,
-0.0607321560382843,
-0.017948897555470467,
0.09861468523740768,
0.0047105178236961365,
-0.016870418563485146,
-0.04580403491854668,
0.08655522763729095,
-0.052611976861953735,
-0.12940941751003265,
0.044319022446870804,
-0.21389912068843842,
-0.18189968168735504,
-0.01621951162815094,
0.0737440437078476,
0.025907833129167557,
0.07358670234680176,
0.023658396676182747,
-0.01594620943069458,
0.07266279309988022,
-0.002135667484253645,
-0.03922203928232193,
-0.12449907511472702,
0.09789172559976578,
-0.08541891723871231,
0.22451646625995636,
-0.05113249272108078,
0.033767033368349075,
0.12520696222782135,
0.04805237799882889,
-0.09938365966081619,
0.03676412254571915,
0.08339355885982513,
-0.1628859043121338,
0.02202500030398369,
0.20174278318881989,
-0.026789559051394463,
0.11296169459819794,
0.023052426055073738,
-0.1522214710712433,
0.0070291683077812195,
-0.07197581976652145,
-0.04018205776810646,
-0.07558760792016983,
-0.005068422295153141,
-0.038223572075366974,
0.12557733058929443,
0.22619622945785522,
-0.0733761116862297,
-0.011573482304811478,
-0.0511484369635582,
0.020737338811159134,
0.07531599700450897,
0.06545042246580124,
-0.025729108601808548,
-0.30105701088905334,
0.015656596049666405,
0.013840238563716412,
-0.016263367608189583,
-0.26692089438438416,
-0.07826614379882812,
0.05216296389698982,
-0.06309262663125992,
-0.06845378875732422,
0.0708208754658699,
0.09046905487775803,
0.04738235846161842,
-0.043570782989263535,
-0.05229482054710388,
-0.06007610633969307,
0.17685922980308533,
-0.19462484121322632,
-0.07474099844694138
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper Large V2
This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3247
- Wer: 13.4709
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 20
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:-------:|
| 0.5388 | 0.49 | 30 | 0.3297 | 12.2434 |
| 0.2858 | 0.98 | 60 | 0.2893 | 23.3419 |
| 0.143 | 1.48 | 90 | 0.2922 | 13.5327 |
| 0.1337 | 1.97 | 120 | 0.2838 | 10.7065 |
| 0.0606 | 2.46 | 150 | 0.2905 | 10.3765 |
| 0.0557 | 2.95 | 180 | 0.2915 | 10.0258 |
| 0.0265 | 3.44 | 210 | 0.3139 | 10.8613 |
| 0.0207 | 3.93 | 240 | 0.3094 | 10.0670 |
| 0.0098 | 4.43 | 270 | 0.3188 | 12.0578 |
| 0.0098 | 4.92 | 300 | 0.3247 | 13.4709 |
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"language": ["nl"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "openai/whisper-large-v2", "model-index": [{"name": "Whisper Large V2", "results": []}]} | automatic-speech-recognition | golesheed/whisper-native-elderly-4-dutch | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"generated_from_trainer",
"nl",
"base_model:openai/whisper-large-v2",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-06T08:15:34+00:00 | [] | [
"nl"
] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #nl #base_model-openai/whisper-large-v2 #license-apache-2.0 #endpoints_compatible #region-us
| Whisper Large V2
================
This model is a fine-tuned version of openai/whisper-large-v2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.3247
* Wer: 13.4709
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 16
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 20
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 20\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #nl #base_model-openai/whisper-large-v2 #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 20\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
74,
116,
4,
38
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #nl #base_model-openai/whisper-large-v2 #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 20\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.13621632754802704,
0.15968677401542664,
-0.001858389237895608,
0.08833666890859604,
0.08782155066728592,
-0.012314315885305405,
0.1681707203388214,
0.12549269199371338,
-0.0416557751595974,
0.10454203933477402,
0.11453307420015335,
0.07736121118068695,
0.04561321437358856,
0.19964902102947235,
-0.06639396399259567,
-0.18399325013160706,
0.06508026272058487,
-0.01929723657667637,
-0.012141751125454903,
0.10182169824838638,
0.06421684473752975,
-0.12256090342998505,
0.06952235847711563,
-0.005073512438684702,
-0.12787765264511108,
-0.04992425814270973,
-0.003603593911975622,
-0.08829036355018616,
0.10478577762842178,
0.011877073906362057,
0.07441696524620056,
0.055234555155038834,
0.0800032690167427,
-0.20251408219337463,
0.008391182869672775,
0.03783120587468147,
0.020442096516489983,
0.06881789863109589,
0.023107007145881653,
0.002746920334175229,
0.017535895109176636,
-0.09417406469583511,
0.06630997359752655,
0.024000227451324463,
-0.09732259809970856,
-0.2184467762708664,
-0.10026310384273529,
0.03434481471776962,
0.08626585453748703,
0.07408887147903442,
-0.013106338679790497,
0.1581668108701706,
-0.012119011022150517,
0.09104708582162857,
0.23101380467414856,
-0.3176605999469757,
-0.04742036387324333,
-0.014271175488829613,
0.026294108480215073,
0.06450144946575165,
-0.07825247943401337,
-0.004888627678155899,
0.050102706998586655,
0.031746502965688705,
0.10774286836385727,
-0.004931247793138027,
-0.06233746558427811,
-0.02886657603085041,
-0.14197440445423126,
-0.05681544914841652,
0.15547265112400055,
0.045779261738061905,
-0.06491699069738388,
-0.09191487729549408,
-0.07131195068359375,
-0.1496184766292572,
-0.04956618696451187,
-0.0026261198800057173,
0.040708139538764954,
-0.04003528878092766,
-0.10001271963119507,
-0.006219570990651846,
-0.06729330122470856,
-0.08073122799396515,
-0.03193339332938194,
0.14468765258789062,
0.03321274742484093,
0.009420378133654594,
-0.014335672371089458,
0.05693294107913971,
-0.0019842665642499924,
-0.16536004841327667,
-0.03401283174753189,
0.033182013779878616,
-0.021878091618418694,
-0.020712248980998993,
-0.03135780245065689,
-0.03883081674575806,
0.06537950783967972,
0.15842470526695251,
-0.04659705609083176,
0.06207548454403877,
-0.0385638028383255,
0.023183785378932953,
-0.09648232907056808,
0.18515583872795105,
-0.044731661677360535,
-0.04666309803724289,
0.029621554538607597,
0.1070999950170517,
0.09432975202798843,
-0.03299562633037567,
-0.09444667398929596,
0.03328503668308258,
0.11510292440652847,
0.05439945310354233,
-0.02153661474585533,
0.05379653349518776,
-0.037912797182798386,
0.004595438949763775,
0.04724494367837906,
-0.1189795732498169,
0.013170532882213593,
0.012636043131351471,
-0.043195948004722595,
-0.055782828480005264,
0.029871907085180283,
0.015721013769507408,
-0.01764751970767975,
0.04835459217429161,
-0.06849675625562668,
-0.011128749698400497,
-0.03830641880631447,
-0.10412869602441788,
0.03652309626340866,
-0.08999405056238174,
-0.004184857476502657,
-0.1197713240981102,
-0.14654004573822021,
-0.010356196202337742,
0.02779223397374153,
-0.03712131455540657,
-0.03228628262877464,
-0.09426835924386978,
-0.11704937368631363,
0.032112982124090195,
-0.02710411697626114,
0.014512862078845501,
-0.07628854364156723,
0.08355537056922913,
0.04701763018965721,
0.09049330651760101,
-0.03954431042075157,
0.01878131553530693,
-0.08150075376033783,
0.04918791353702545,
-0.17111089825630188,
0.07168842107057571,
-0.08916273713111877,
0.07433856278657913,
-0.11224762350320816,
-0.07275255769491196,
0.043903231620788574,
-0.023215988650918007,
0.102077417075634,
0.11524751782417297,
-0.19717739522457123,
-0.04972013831138611,
0.21730080246925354,
-0.1178780272603035,
-0.15726397931575775,
0.15235035121440887,
-0.01904238387942314,
-0.018581777811050415,
0.06867604702711105,
0.24456007778644562,
0.06123080849647522,
-0.1375725120306015,
-0.056367188692092896,
-0.0261027030646801,
0.07121604681015015,
-0.05355974659323692,
0.07951906323432922,
-0.004325232934206724,
0.04291364923119545,
0.010138435289263725,
-0.01653929427266121,
0.04194147512316704,
-0.07492784410715103,
-0.09569421410560608,
-0.04744570702314377,
-0.11836248636245728,
0.01733187958598137,
0.002287462819367647,
0.02962525747716427,
-0.12062839418649673,
-0.08092740178108215,
0.0054471115581691265,
0.11621628701686859,
-0.09865685552358627,
0.026550618931651115,
-0.1331365555524826,
0.11432939022779465,
-0.07463306933641434,
-0.010850919410586357,
-0.1483161598443985,
-0.018835563212633133,
0.044661860913038254,
-0.03598916903138161,
0.02147132158279419,
-0.0885128602385521,
0.07264025509357452,
0.07467518001794815,
-0.02658570371568203,
-0.04248100146651268,
-0.0021844636648893356,
0.011659912765026093,
-0.08558013290166855,
-0.20805923640727997,
-0.02916303463280201,
-0.05072628706693649,
0.13512323796749115,
-0.16510847210884094,
0.026289111003279686,
0.036941152065992355,
0.10755885392427444,
0.050888173282146454,
-0.028413690626621246,
0.012378765270113945,
0.0576338954269886,
-0.021883849054574966,
-0.07558631896972656,
0.028089145198464394,
0.03416816145181656,
-0.09576820582151413,
0.017973311245441437,
-0.1792900711297989,
0.13350984454154968,
0.13961872458457947,
0.05354544147849083,
-0.03920949250459671,
0.026662219315767288,
-0.03780747950077057,
-0.03187672048807144,
-0.026316704228520393,
0.004363284446299076,
0.1465822458267212,
0.0021496822591871023,
0.12176033109426498,
-0.09753876179456711,
-0.02674207277595997,
0.045568548142910004,
-0.03195665404200554,
-0.012703437358140945,
0.09416963160037994,
0.021390480920672417,
-0.07712744921445847,
0.11574635654687881,
0.1106874868273735,
-0.07978503406047821,
0.11232184618711472,
-0.07221370190382004,
-0.05422946438193321,
-0.02878420613706112,
0.027434226125478745,
0.043337490409612656,
0.12183227390050888,
-0.09068036079406738,
-0.01438656821846962,
0.02515447698533535,
0.012049616314470768,
0.006123725324869156,
-0.1886918991804123,
0.005767221562564373,
0.016797926276922226,
-0.09008780866861343,
-0.025112684816122055,
-0.006492929067462683,
-0.011711121536791325,
0.09337366372346878,
-0.00300964480265975,
-0.10617176443338394,
0.014498971402645111,
-0.017466768622398376,
-0.07522863149642944,
0.17250363528728485,
-0.09990600496530533,
-0.15192309021949768,
-0.11562493443489075,
-0.048458099365234375,
-0.04277130961418152,
0.020420284941792488,
0.06840915232896805,
-0.06886673718690872,
-0.04794713854789734,
-0.12919248640537262,
-0.04088391363620758,
0.06478952616453171,
0.04791481792926788,
0.08473362773656845,
-0.0035977375227957964,
0.0837816372513771,
-0.10559368878602982,
-0.008524681441485882,
-0.028159063309431076,
-0.01048053614795208,
0.014065103605389595,
0.03254552185535431,
0.11960489302873611,
0.1376330703496933,
-0.000840022403281182,
0.019899478182196617,
-0.028172969818115234,
0.2301720380783081,
-0.07224448770284653,
-0.031108040362596512,
0.12259376049041748,
-0.02640325389802456,
0.051638029515743256,
0.17324084043502808,
0.030002977699041367,
-0.11868065595626831,
-0.00007048338738968596,
-0.023971926420927048,
-0.04381057620048523,
-0.21067218482494354,
-0.06343332678079605,
-0.038883596658706665,
0.04803768917918205,
0.07741225510835648,
0.03194250538945198,
0.01576736941933632,
0.0287628173828125,
0.010801567696034908,
0.03336828202009201,
-0.000778556102886796,
0.07745110988616943,
0.10073007643222809,
0.053871072828769684,
0.11427044123411179,
-0.04864861071109772,
-0.03424227237701416,
0.0309415552765131,
0.008827924728393555,
0.21411235630512238,
-0.006806745659559965,
0.1936895102262497,
0.031086552888154984,
0.14399775862693787,
0.03124992921948433,
0.05868235230445862,
-0.004321998450905085,
-0.0014536561211571097,
0.00011419778456911445,
-0.06801947951316833,
-0.049581386148929596,
0.026712197810411453,
-0.027490390464663506,
0.04718725383281708,
-0.09096401929855347,
0.06585752964019775,
0.06531890481710434,
0.291382759809494,
0.08233320713043213,
-0.35608431696891785,
-0.10410318523645401,
0.013939191587269306,
-0.04568859934806824,
-0.010707268491387367,
0.04109747335314751,
0.1676306277513504,
-0.035800375044345856,
0.05957348644733429,
-0.04908604547381401,
0.07021871209144592,
-0.07418491691350937,
0.031817398965358734,
0.02570035308599472,
0.0855240598320961,
0.0018178020836785436,
0.023594986647367477,
-0.2272135466337204,
0.28047269582748413,
0.012436628341674805,
0.10276336967945099,
-0.0487116314470768,
0.0020353891886770725,
0.02182648703455925,
0.016720861196517944,
0.10008277744054794,
-0.015368645079433918,
-0.12311194837093353,
-0.1574331372976303,
-0.1477552056312561,
0.045148614794015884,
0.09810465574264526,
0.0223187617957592,
0.11414927989244461,
-0.014110207557678223,
-0.032044459134340286,
0.04537680372595787,
-0.06469541043043137,
-0.053272537887096405,
-0.07595355063676834,
0.006568394135683775,
0.11694630980491638,
0.004027381539344788,
-0.07790198922157288,
-0.09135306626558304,
-0.09162882715463638,
0.11549058556556702,
-0.04482784494757652,
-0.03849310800433159,
-0.09641735255718231,
0.003951740451157093,
0.11544102430343628,
-0.08045060932636261,
0.044089674949645996,
0.018454937264323235,
0.10842259973287582,
0.01175297424197197,
-0.05085345357656479,
0.10011888295412064,
-0.08017635345458984,
-0.1976318061351776,
-0.04278148338198662,
0.14413614571094513,
0.007582133635878563,
0.0570187009871006,
0.016380207613110542,
0.03411366418004036,
0.002453245222568512,
-0.0755670964717865,
0.03509826213121414,
0.05489535257220268,
0.017490698024630547,
0.005914448760449886,
0.014068059623241425,
-0.05549638345837593,
-0.07103073596954346,
-0.026736579835414886,
0.16581539809703827,
0.29358914494514465,
-0.07713106274604797,
0.07019561529159546,
0.10554484277963638,
-0.033897191286087036,
-0.20164474844932556,
-0.01919488050043583,
0.044733766466379166,
0.014007529243826866,
-0.03216458484530449,
-0.13453763723373413,
0.07363015413284302,
0.06231628358364105,
-0.048025429248809814,
0.07798503339290619,
-0.29027286171913147,
-0.14117367565631866,
0.11903156340122223,
0.10698550939559937,
0.09159891307353973,
-0.13823311030864716,
-0.06350408494472504,
-0.03951125591993332,
-0.10792451351881027,
0.0806211605668068,
-0.13480418920516968,
0.11504227668046951,
0.011054912582039833,
0.054707806557416916,
0.009646838530898094,
-0.06326767057180405,
0.1307542473077774,
0.011582973413169384,
0.07547330111265182,
-0.04929868131875992,
0.01807730831205845,
0.026336537674069405,
-0.07765660434961319,
0.06403639912605286,
-0.1037367582321167,
0.06644802540540695,
-0.01571423001587391,
-0.02818179689347744,
-0.05265707150101662,
0.0024104963522404432,
-0.001534575829282403,
-0.030629007145762444,
-0.024864450097084045,
0.017817918211221695,
0.07605646550655365,
0.003878478892147541,
0.12219811975955963,
0.003997919615358114,
0.10227310657501221,
0.14183922111988068,
0.11386533826589584,
-0.08582790940999985,
-0.0023103870917111635,
0.017696576192975044,
-0.05048684403300285,
0.056129369884729385,
-0.12267884612083435,
0.05080711469054222,
0.1187099888920784,
0.02622661180794239,
0.11883791536092758,
0.05286730080842972,
-0.05940454453229904,
0.034165505319833755,
0.05844949930906296,
-0.14980162680149078,
-0.1553071290254593,
0.023392148315906525,
0.03544645383954048,
-0.1201973706483841,
0.08259966224431992,
0.14752541482448578,
-0.0755700096487999,
0.0036739788483828306,
-0.01621028408408165,
0.031722839921712875,
-0.02674170583486557,
0.2017463594675064,
0.04023810103535652,
0.0588664673268795,
-0.11580227315425873,
0.0918835923075676,
0.04340014234185219,
-0.09641899168491364,
0.07375139743089676,
0.05963040515780449,
-0.11412306874990463,
-0.029042372480034828,
0.012517434544861317,
0.1298513263463974,
0.023425843566656113,
-0.0739997997879982,
-0.13419552147388458,
-0.11811395734548569,
0.07012086361646652,
0.2163677215576172,
0.06164073571562767,
0.03168521821498871,
-0.017447616904973984,
0.008418331854045391,
-0.11339852213859558,
0.11061214655637741,
0.04519093409180641,
0.06823432445526123,
-0.14646613597869873,
0.11587050557136536,
-0.015727465972304344,
0.02495487593114376,
-0.02533199079334736,
0.026403261348605156,
-0.11744826287031174,
0.01147575955837965,
-0.14627821743488312,
0.05394823104143143,
-0.05587800592184067,
0.004396585747599602,
0.00706110754981637,
-0.056114312261343,
-0.0706772431731224,
0.034763094037771225,
-0.09286361187696457,
-0.030004320666193962,
0.0049022468738257885,
0.03647990897297859,
-0.13249318301677704,
-0.03307902812957764,
0.018925711512565613,
-0.09960555285215378,
0.1019238755106926,
0.05885067582130432,
-0.018056461587548256,
0.04574736952781677,
-0.1150609701871872,
-0.03010399267077446,
0.07337243109941483,
0.014473870396614075,
0.05624442547559738,
-0.12246954441070557,
-0.03631245344877243,
0.01771981082856655,
0.022298045456409454,
0.01760168746113777,
0.11799582093954086,
-0.09618406742811203,
0.01204153336584568,
-0.022656630724668503,
-0.01208792719990015,
-0.05695350468158722,
0.01780855469405651,
0.11104132980108261,
0.022894855588674545,
0.15558363497257233,
-0.09993931651115417,
0.008565147407352924,
-0.1832948923110962,
-0.0005412423051893711,
-0.019095033407211304,
-0.11106836795806885,
-0.11574822664260864,
-0.0049544572830200195,
0.08544056117534637,
-0.07203175127506256,
0.11254741251468658,
-0.05579017847776413,
0.01858634501695633,
0.02793101966381073,
-0.052619196474552155,
-0.03907065838575363,
0.05091674625873566,
0.20321890711784363,
0.04325638338923454,
-0.03485757112503052,
0.06155044957995415,
-0.014558817259967327,
0.09744757413864136,
0.09995421022176743,
0.15737716853618622,
0.15704390406608582,
0.07030757516622543,
0.12281396985054016,
0.08372491598129272,
-0.05185065045952797,
-0.15671880543231964,
0.05434349551796913,
-0.07844970375299454,
0.11937134712934494,
-0.003698879387229681,
0.19599823653697968,
0.10510700941085815,
-0.12325511127710342,
0.017096078023314476,
-0.04304680973291397,
-0.08304844051599503,
-0.10995044559240341,
-0.06291265785694122,
-0.10993708670139313,
-0.13319607079029083,
0.0006502840551547706,
-0.11679433286190033,
0.02228342927992344,
0.08178167790174484,
0.02289886213839054,
0.019076641649007797,
0.1472093164920807,
-0.008880493231117725,
0.05105264112353325,
0.07077731937170029,
-0.009628964588046074,
-0.05563129484653473,
-0.016096483916044235,
-0.10109423100948334,
0.04725433886051178,
0.025623423978686333,
0.05894210562109947,
-0.009952942840754986,
-0.035064611583948135,
0.07017426937818527,
-0.02025395818054676,
-0.11677761375904083,
0.014983261935412884,
0.00980759970843792,
0.056707728654146194,
0.029363341629505157,
0.05678620934486389,
-0.013942847959697247,
0.01449260301887989,
0.20588436722755432,
-0.09717857837677002,
-0.11521829664707184,
-0.14177344739437103,
0.18006600439548492,
-0.013315352611243725,
-0.014675081707537174,
0.014569193124771118,
-0.08950605243444443,
-0.02605634182691574,
0.17809627950191498,
0.19374603033065796,
-0.05199835076928139,
0.003235545242205262,
-0.03814956545829773,
-0.002785347169265151,
-0.07368181645870209,
0.0733838677406311,
0.12397715449333191,
0.07091996818780899,
-0.05419512465596199,
-0.05520976334810257,
-0.03642336651682854,
-0.030847154557704926,
-0.0404658168554306,
0.03606013208627701,
-0.028954196721315384,
-0.006494378205388784,
-0.04469893127679825,
0.057992979884147644,
-0.09258225560188293,
-0.10527916252613068,
-0.00040223964606411755,
-0.20833276212215424,
-0.1687675267457962,
0.00417300732806325,
0.07911945879459381,
0.03212394937872887,
0.028318211436271667,
-0.012253234162926674,
0.00021580942848231643,
0.08203965425491333,
-0.029652435332536697,
-0.06635201722383499,
-0.04245767742395401,
0.0534469336271286,
-0.09109541773796082,
0.21040211617946625,
-0.019157133996486664,
0.06370812654495239,
0.12168209254741669,
0.06113966926932335,
-0.11188666522502899,
0.08761792629957199,
0.06347261369228363,
-0.10575605183839798,
0.029891571030020714,
0.14618797600269318,
-0.06277395039796829,
0.13370433449745178,
0.05734676495194435,
-0.1043623760342598,
-0.03006412647664547,
-0.02347303181886673,
-0.03269651532173157,
-0.061209358274936676,
-0.05205981805920601,
-0.053864605724811554,
0.1346656084060669,
0.1382024586200714,
-0.07198037952184677,
0.003933044616132975,
-0.009475797414779663,
0.024455105885863304,
0.041742440313100815,
0.0062119378708302975,
-0.03529017046093941,
-0.2672055959701538,
0.009498773142695427,
0.004874990321695805,
0.017476601526141167,
-0.2497686743736267,
-0.07322251796722412,
-0.007318982854485512,
-0.031901322305202484,
-0.07763360440731049,
0.09062253683805466,
0.11694224178791046,
0.03633561730384827,
-0.06715060025453568,
-0.022874874994158745,
-0.031744956970214844,
0.1601707637310028,
-0.15446433424949646,
-0.10832486301660538
] |
null | null | transformers | # IUPAC2SMILES-canonical-base
IUPAC2SMILES-canonical-base was designed to accurately translate IUPAC chemical names to SMILES.
## Model Details
### Model Description
IUPAC2SMILES-canonical-base is based on the MT5 model with optimizations in implementing different tokenizers for the encoder and decoder.
- **Developed by:** Knowladgator Engineering
- **Model type:** Encoder-Decoder with attention mechanism
- **Language(s) (NLP):** SMILES, IUPAC (English)
- **License:** Apache License 2.0
### Model Sources
- **Paper:** coming soon
- **Demo:** [ChemicalConverters](https://huggingface.co/spaces/knowledgator/ChemicalConverters)
## Quickstart
Firstly, install the library:
```commandline
pip install chemical-converters
```
### IUPAC to SMILES
#### To perform simple translation, follow the example:
```python
from chemicalconverters import NamesConverter
converter = NamesConverter(model_name="knowledgator/IUPAC2SMILES-canonical-base")
print(converter.iupac_to_smiles('ethanol'))
print(converter.iupac_to_smiles(['ethanol', 'ethanol', 'ethanol']))
```
```text
['CCO']
['CCO', 'CCO', 'CCO']
```
#### Processing in batches:
```python
from chemicalconverters import NamesConverter
converter = NamesConverter(model_name="knowledgator/IUPAC2SMILES-canonical-base")
print(converter.iupac_to_smiles(["buta-1,3-diene" for _ in range(10)], num_beams=1,
process_in_batch=True, batch_size=1000))
```
```text
['<SYST>C=CC=C', '<SYST>C=CC=C'...]
```
Our models also predict IUPAC styles from the table:
| Style Token | Description |
|-------------|----------------------------------------------------------------------------------------------------|
| `<BASE>` | The most known name of the substance, sometimes is the mixture of traditional and systematic style |
| `<SYST>` | The totally systematic style without trivial names |
| `<TRAD>` | The style is based on trivial names of the parts of substances |
## Bias, Risks, and Limitations
This model has limited accuracy in processing large molecules and currently, doesn't support isomeric and isotopic SMILES.
### Training Procedure
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
The model was trained on 100M examples of SMILES-IUPAC pairs with lr=0.00001, batch_size=512 for 2 epochs.
## Evaluation
| Model | Accuracy | BLEU-4 score | Size(MB) |
|-------------------------------------|---------|------------------|----------|
| IUPAC2SMILES-canonical-small |88.9% |0.966 |23 |
| IUPAC2SMILES-canonical-base |93.7% |0.974 |180 |
| STOUT V2.0\* |68.47% |0.92 |128 |
*According to the original paper https://jcheminf.biomedcentral.com/articles/10.1186/s13321-021-00512-4
## Citation
Coming soon.
## Model Card Authors
[Mykhailo Shtopko](https://huggingface.co/BioMike)
## Model Card Contact
[email protected] | {"license": "apache-2.0", "tags": ["chemistry", "biology", "medical", "smiles", "iupac", "text-generation-inference"], "metrics": ["accuracy", "bleu"], "pipeline_tag": "text2text-generation", "widget": [{"text": "ethanol", "example_title": "CCO"}]} | text2text-generation | knowledgator/IUPAC2SMILES-canonical-base | [
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"chemistry",
"biology",
"medical",
"smiles",
"iupac",
"text-generation-inference",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | 2024-02-06T08:15:49+00:00 | [] | [] | TAGS
#transformers #pytorch #mt5 #text2text-generation #chemistry #biology #medical #smiles #iupac #text-generation-inference #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
| IUPAC2SMILES-canonical-base
===========================
IUPAC2SMILES-canonical-base was designed to accurately translate IUPAC chemical names to SMILES.
Model Details
-------------
### Model Description
IUPAC2SMILES-canonical-base is based on the MT5 model with optimizations in implementing different tokenizers for the encoder and decoder.
* Developed by: Knowladgator Engineering
* Model type: Encoder-Decoder with attention mechanism
* Language(s) (NLP): SMILES, IUPAC (English)
* License: Apache License 2.0
### Model Sources
* Paper: coming soon
* Demo: ChemicalConverters
Quickstart
----------
Firstly, install the library:
### IUPAC to SMILES
#### To perform simple translation, follow the example:
#### Processing in batches:
Our models also predict IUPAC styles from the table:
Bias, Risks, and Limitations
----------------------------
This model has limited accuracy in processing large molecules and currently, doesn't support isomeric and isotopic SMILES.
### Training Procedure
The model was trained on 100M examples of SMILES-IUPAC pairs with lr=0.00001, batch\_size=512 for 2 epochs.
Evaluation
----------
Coming soon.
Model Card Authors
------------------
Mykhailo Shtopko
Model Card Contact
------------------
info@URL
| [
"### Model Description\n\n\nIUPAC2SMILES-canonical-base is based on the MT5 model with optimizations in implementing different tokenizers for the encoder and decoder.\n\n\n* Developed by: Knowladgator Engineering\n* Model type: Encoder-Decoder with attention mechanism\n* Language(s) (NLP): SMILES, IUPAC (English)\n* License: Apache License 2.0",
"### Model Sources\n\n\n* Paper: coming soon\n* Demo: ChemicalConverters\n\n\nQuickstart\n----------\n\n\nFirstly, install the library:",
"### IUPAC to SMILES",
"#### To perform simple translation, follow the example:",
"#### Processing in batches:\n\n\nOur models also predict IUPAC styles from the table:\n\n\n\nBias, Risks, and Limitations\n----------------------------\n\n\nThis model has limited accuracy in processing large molecules and currently, doesn't support isomeric and isotopic SMILES.",
"### Training Procedure\n\n\nThe model was trained on 100M examples of SMILES-IUPAC pairs with lr=0.00001, batch\\_size=512 for 2 epochs.\n\n\nEvaluation\n----------\n\n\n\nComing soon.\n\n\nModel Card Authors\n------------------\n\n\nMykhailo Shtopko\n\n\nModel Card Contact\n------------------\n\n\ninfo@URL"
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #chemistry #biology #medical #smiles #iupac #text-generation-inference #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"### Model Description\n\n\nIUPAC2SMILES-canonical-base is based on the MT5 model with optimizations in implementing different tokenizers for the encoder and decoder.\n\n\n* Developed by: Knowladgator Engineering\n* Model type: Encoder-Decoder with attention mechanism\n* Language(s) (NLP): SMILES, IUPAC (English)\n* License: Apache License 2.0",
"### Model Sources\n\n\n* Paper: coming soon\n* Demo: ChemicalConverters\n\n\nQuickstart\n----------\n\n\nFirstly, install the library:",
"### IUPAC to SMILES",
"#### To perform simple translation, follow the example:",
"#### Processing in batches:\n\n\nOur models also predict IUPAC styles from the table:\n\n\n\nBias, Risks, and Limitations\n----------------------------\n\n\nThis model has limited accuracy in processing large molecules and currently, doesn't support isomeric and isotopic SMILES.",
"### Training Procedure\n\n\nThe model was trained on 100M examples of SMILES-IUPAC pairs with lr=0.00001, batch\\_size=512 for 2 epochs.\n\n\nEvaluation\n----------\n\n\n\nComing soon.\n\n\nModel Card Authors\n------------------\n\n\nMykhailo Shtopko\n\n\nModel Card Contact\n------------------\n\n\ninfo@URL"
] | [
78,
94,
30,
9,
11,
63,
73
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #text2text-generation #chemistry #biology #medical #smiles #iupac #text-generation-inference #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n### Model Description\n\n\nIUPAC2SMILES-canonical-base is based on the MT5 model with optimizations in implementing different tokenizers for the encoder and decoder.\n\n\n* Developed by: Knowladgator Engineering\n* Model type: Encoder-Decoder with attention mechanism\n* Language(s) (NLP): SMILES, IUPAC (English)\n* License: Apache License 2.0### Model Sources\n\n\n* Paper: coming soon\n* Demo: ChemicalConverters\n\n\nQuickstart\n----------\n\n\nFirstly, install the library:### IUPAC to SMILES#### To perform simple translation, follow the example:#### Processing in batches:\n\n\nOur models also predict IUPAC styles from the table:\n\n\n\nBias, Risks, and Limitations\n----------------------------\n\n\nThis model has limited accuracy in processing large molecules and currently, doesn't support isomeric and isotopic SMILES.### Training Procedure\n\n\nThe model was trained on 100M examples of SMILES-IUPAC pairs with lr=0.00001, batch\\_size=512 for 2 epochs.\n\n\nEvaluation\n----------\n\n\n\nComing soon.\n\n\nModel Card Authors\n------------------\n\n\nMykhailo Shtopko\n\n\nModel Card Contact\n------------------\n\n\ninfo@URL"
] | [
-0.07426460832357407,
0.1717124730348587,
-0.0030647756066173315,
-0.015530804172158241,
0.045839592814445496,
0.013356316834688187,
0.12014619261026382,
0.14481109380722046,
0.03867899626493454,
0.14497612416744232,
0.06625079363584518,
0.03162157163023949,
0.10001114010810852,
0.18725384771823883,
0.03873114660382271,
-0.2249518632888794,
0.07299216836690903,
-0.10684007406234741,
-0.07457195222377777,
0.10600138455629349,
0.10340218245983124,
-0.05776910483837128,
0.06953675299882889,
-0.01883113570511341,
-0.0257231704890728,
-0.012817972339689732,
-0.06813885271549225,
-0.08043279498815536,
0.043540846556425095,
0.01134704053401947,
0.031750500202178955,
0.021995581686496735,
0.060616374015808105,
-0.27903956174850464,
0.014493661932647228,
0.047046586871147156,
0.04771025478839874,
0.06065120920538902,
0.11986016482114792,
-0.04959113895893097,
0.1418299376964569,
-0.15403224527835846,
0.07260207086801529,
0.052054423838853836,
-0.07873713225126266,
-0.09740789979696274,
-0.09875985234975815,
0.11892946809530258,
0.09466729313135147,
0.02829739823937416,
-0.027753394097089767,
0.09093640744686127,
-0.007171482779085636,
0.03526919707655907,
0.17126679420471191,
-0.17533569037914276,
-0.012157545424997807,
0.05221683159470558,
0.04124580696225166,
0.07179079204797745,
-0.030795246362686157,
0.020465098321437836,
0.023304034024477005,
-0.01059021707624197,
0.0382574237883091,
-0.009382707066833973,
0.10593915730714798,
-0.008435838855803013,
-0.15075090527534485,
-0.09779518097639084,
0.14023138582706451,
0.006420504301786423,
-0.07029847055673599,
-0.17690037190914154,
-0.033693037927150726,
-0.053088732063770294,
-0.06350712478160858,
-0.05873541533946991,
0.06043533608317375,
0.004033134318888187,
0.06806322932243347,
-0.015314364805817604,
-0.0697479397058487,
-0.04359806701540947,
0.0637955293059349,
0.07369273900985718,
0.05535604804754257,
0.007158761378377676,
0.016654182225465775,
0.07825654000043869,
-0.015398683957755566,
-0.09822370111942291,
-0.06039711460471153,
-0.06281279772520065,
-0.06690005213022232,
-0.024589095264673233,
-0.00657852366566658,
-0.008940111845731735,
0.08087774366140366,
0.15118113160133362,
-0.07775828242301941,
0.05975835397839546,
0.06473788619041443,
0.007179804611951113,
0.04864955320954323,
0.09477313607931137,
-0.053270574659109116,
-0.07868842035531998,
-0.036945220082998276,
0.07332217693328857,
-0.014897269196808338,
0.017534887418150902,
-0.00358585431240499,
0.0367492213845253,
0.07469220459461212,
0.08352750539779663,
0.007771277334541082,
0.030443981289863586,
-0.05533740296959877,
-0.03947797790169716,
0.09644115716218948,
-0.14326009154319763,
0.01094536017626524,
0.040300317108631134,
-0.046282753348350525,
0.08934129029512405,
0.019441785290837288,
-0.022161083295941353,
-0.08074073493480682,
0.021722057834267616,
-0.06580758839845657,
-0.030285941436886787,
-0.10908106714487076,
-0.07232437282800674,
0.030308105051517487,
0.045645784586668015,
-0.05998145043849945,
-0.08368437737226486,
-0.1416332721710205,
-0.07398567348718643,
0.053793102502822876,
-0.0695667639374733,
0.0035168719477951527,
-0.03689839690923691,
-0.028245888650417328,
0.05384979769587517,
0.02918943390250206,
0.059747930616140366,
-0.05594922974705696,
0.01007844042032957,
-0.04757780581712723,
0.11193952709436417,
0.14192579686641693,
0.045191071927547455,
-0.11731039732694626,
0.07699345052242279,
-0.17241084575653076,
0.08173301815986633,
-0.1184268519282341,
0.005516589619219303,
-0.17262718081474304,
-0.07997380197048187,
-0.010748757049441338,
0.007980076596140862,
0.03310896456241608,
0.18537308275699615,
-0.1514449119567871,
-0.055231254547834396,
0.19982290267944336,
-0.10292952507734299,
-0.05620335787534714,
0.08007869124412537,
0.006346381269395351,
0.04463709145784378,
0.02794928289949894,
0.1074696034193039,
0.043624937534332275,
-0.1334465742111206,
-0.04065145552158356,
-0.001942514325492084,
0.017401060089468956,
0.08506408333778381,
0.09189945459365845,
-0.06092490628361702,
0.010722583159804344,
0.02827194333076477,
-0.0815584659576416,
-0.01911691203713417,
-0.04757875204086304,
-0.06282258033752441,
0.0010575124761089683,
-0.05116984620690346,
-0.020898807793855667,
-0.039723265916109085,
0.00911399070173502,
-0.04731842502951622,
-0.10543157905340195,
-0.010805905796587467,
0.12004470080137253,
-0.04603482782840729,
-0.0006838577101007104,
-0.05888713151216507,
0.09088417142629623,
-0.003420293563976884,
0.00937923975288868,
-0.14154094457626343,
0.0034954205621033907,
0.05894331634044647,
-0.17805205285549164,
0.021621622145175934,
-0.08034873008728027,
0.023287678137421608,
0.06819859147071838,
0.014059041626751423,
-0.02359696850180626,
0.006924229208379984,
0.010572620667517185,
-0.05098980292677879,
-0.18621209263801575,
-0.023807302117347717,
-0.0667770653963089,
0.15869386494159698,
-0.1695835292339325,
0.02432992123067379,
0.03575452044606209,
0.0932837426662445,
0.010016593150794506,
-0.0448254831135273,
0.08371781557798386,
-0.0038255462422966957,
-0.008412465453147888,
-0.09024374186992645,
0.050208207219839096,
-0.027235155925154686,
-0.012452533468604088,
0.07165262848138809,
-0.1419305056333542,
-0.16437390446662903,
0.09889446198940277,
0.07375890016555786,
-0.10146839916706085,
-0.0916977971792221,
-0.04371056333184242,
-0.028004851192235947,
-0.01958240382373333,
-0.014490257948637009,
0.06117798388004303,
0.04510900750756264,
0.0938582494854927,
-0.0803954154253006,
-0.02981787919998169,
-0.008991861715912819,
-0.04070429876446724,
-0.0892692431807518,
0.09823646396398544,
0.09096803516149521,
-0.1151520237326622,
0.07977529615163803,
0.09418398141860962,
0.06503196805715561,
0.14965565502643585,
0.0029852131847292185,
-0.12888552248477936,
0.0010563898831605911,
0.02919437177479267,
0.03118950128555298,
0.08145933598279953,
-0.06314203888177872,
0.00467307772487402,
0.06261596828699112,
0.00312103726901114,
0.03838435932993889,
-0.10101684182882309,
0.06377420574426651,
0.029899097979068756,
-0.04148576408624649,
0.0321807824075222,
-0.02418132871389389,
0.003393955994397402,
0.08649415522813797,
0.03169345483183861,
0.021526196971535683,
-0.05332042649388313,
-0.04205213859677315,
-0.12466991692781448,
0.21149788796901703,
-0.11652840673923492,
-0.22334565222263336,
-0.14698606729507446,
0.013015513308346272,
0.025533821433782578,
-0.0018371718470007181,
-0.0022418496664613485,
-0.06681571900844574,
-0.08261925727128983,
-0.11758896708488464,
0.017863813787698746,
0.0167706198990345,
-0.029115034267306328,
-0.006963308900594711,
0.02200600877404213,
0.03048143908381462,
-0.1368180513381958,
-0.018399519845843315,
0.026709118857979774,
-0.016698379069566727,
0.05247962474822998,
-0.03262782841920853,
0.061209484934806824,
0.1893886923789978,
0.0287006925791502,
0.007804034277796745,
-0.011334232985973358,
0.1890333592891693,
-0.0864429697394371,
0.06364916265010834,
0.1234297901391983,
0.017430255189538002,
0.075093112885952,
0.10081534832715988,
0.00824407022446394,
-0.09171290695667267,
0.05582176148891449,
-0.016909416764974594,
-0.041115567088127136,
-0.2232796549797058,
-0.0974268987774849,
-0.02681819535791874,
0.02892487868666649,
0.07892875373363495,
0.04019475355744362,
0.029916973784565926,
0.05979413539171219,
-0.035728245973587036,
-0.026383033022284508,
-0.009889813140034676,
0.1234225332736969,
0.026479186490178108,
0.01265974622219801,
0.05612252652645111,
-0.02866842783987522,
0.021777531132102013,
0.09593827277421951,
0.013968301936984062,
0.18834121525287628,
0.009903935715556145,
0.17330330610275269,
0.0983201116323471,
0.05825229734182358,
0.0512550063431263,
0.07048434019088745,
-0.0020889255683869123,
0.04926101863384247,
-0.041376933455467224,
-0.10754649341106415,
-0.017705345526337624,
0.11020118743181229,
-0.01955335959792137,
-0.018765980377793312,
-0.029538515955209732,
0.08265721797943115,
0.0393066368997097,
0.26463764905929565,
0.07576636224985123,
-0.22626234591007233,
-0.03893919289112091,
0.07193166017532349,
-0.04326882213354111,
-0.05214657634496689,
0.019437989220023155,
0.034482285380363464,
-0.13569782674312592,
0.02554262802004814,
-0.03885926306247711,
0.0814153328537941,
-0.05543607473373413,
-0.011884711682796478,
0.034104786813259125,
0.10355326533317566,
-0.028237292543053627,
0.0877036526799202,
-0.17348189651966095,
0.17961911857128143,
-0.007147388067096472,
0.06480223685503006,
-0.08604800701141357,
0.06442078202962875,
0.051810372620821,
-0.004862913396209478,
0.14382769167423248,
0.03600141778588295,
-0.17167457938194275,
-0.09449630230665207,
-0.13935647904872894,
-0.025785142555832863,
0.10752590000629425,
-0.008295928128063679,
0.1275608241558075,
-0.02484801411628723,
-0.03180354833602905,
-0.04284035786986351,
-0.046716537326574326,
-0.14644348621368408,
-0.13939188420772552,
0.06974267959594727,
-0.10571584105491638,
0.06725897639989853,
-0.0865388736128807,
-0.043121371418237686,
-0.04516727849841118,
0.13873469829559326,
-0.10414335131645203,
-0.08983375877141953,
-0.14957337081432343,
-0.0009267244022339582,
0.1484006941318512,
-0.06887940317392349,
0.023083822801709175,
-0.009998325258493423,
0.12592779099941254,
-0.02776140533387661,
-0.08188508450984955,
0.07602114975452423,
-0.12749283015727997,
-0.22184129059314728,
-0.051733873784542084,
0.13411514461040497,
0.040780749171972275,
0.06917841732501984,
0.012936653569340706,
0.051359422504901886,
-0.00576031394302845,
-0.10254038125276566,
0.05078504979610443,
0.13657592236995697,
0.07366462796926498,
0.07444125413894653,
-0.06993299722671509,
-0.1240384429693222,
-0.10526260733604431,
-0.06258345395326614,
0.08378717303276062,
0.255996435880661,
-0.04667113721370697,
0.15114283561706543,
0.1748708188533783,
-0.12724921107292175,
-0.1602223962545395,
-0.023165149614214897,
0.057706136256456375,
0.011905956082046032,
0.008016614243388176,
-0.20668749511241913,
0.08293943852186203,
0.06901432573795319,
-0.026588432490825653,
0.046231623739004135,
-0.31176191568374634,
-0.14121346175670624,
0.06561728566884995,
0.07145947217941284,
-0.029547160491347313,
-0.1421375274658203,
-0.06255980581045151,
-0.03549553081393242,
-0.12768714129924774,
0.07672069221735,
0.03098624385893345,
0.0697057768702507,
-0.010916461236774921,
0.03468400239944458,
0.016673672944307327,
-0.0648706704378128,
0.14904972910881042,
-0.00259297713637352,
0.02422836422920227,
-0.0425981841981411,
-0.003513459814712405,
0.0038991759065538645,
-0.03460721671581268,
0.12955839931964874,
-0.018989594653248787,
0.030502330511808395,
-0.14718738198280334,
-0.037386078387498856,
-0.04041066765785217,
0.03189944103360176,
-0.05541672930121422,
-0.07701502740383148,
-0.0718669518828392,
0.10832004994153976,
0.10618820786476135,
-0.034138377755880356,
-0.05541475862264633,
-0.04625498875975609,
0.01650705188512802,
0.16045869886875153,
0.10314874351024628,
0.034029893577098846,
-0.08109991252422333,
0.019562866538763046,
-0.0031977323815226555,
0.022236332297325134,
-0.13542525470256805,
0.011685804463922977,
0.07826802134513855,
0.041200216859579086,
0.14043769240379333,
-0.0037490404210984707,
-0.13019511103630066,
-0.018800387158989906,
0.06909441947937012,
-0.10968124866485596,
-0.12883131206035614,
0.017071962356567383,
0.08743438124656677,
-0.07230549305677414,
-0.10142401605844498,
0.08854826539754868,
-0.07042169570922852,
-0.030243024230003357,
-0.025583362206816673,
0.11061622947454453,
0.010028000921010971,
0.15052978694438934,
0.053691864013671875,
0.0754687488079071,
-0.051985468715429306,
0.10263542085886002,
0.0755453109741211,
-0.12790533900260925,
0.050609342753887177,
0.12676462531089783,
-0.0924890860915184,
-0.03593754395842552,
0.04803657904267311,
0.08982904255390167,
-0.010350793600082397,
-0.06885728240013123,
-0.04535720869898796,
-0.08552965521812439,
0.07994011789560318,
0.08868742734193802,
0.04078791290521622,
-0.005251438822597265,
-0.029295016080141068,
0.003709176555275917,
-0.06398672610521317,
0.1088327169418335,
0.04526032134890556,
0.02382367104291916,
-0.040762562304735184,
0.09406805783510208,
-0.011812621727585793,
-0.01680079475045204,
-0.03315645083785057,
-0.03311082720756531,
-0.10144172608852386,
-0.029002394527196884,
-0.12520384788513184,
0.040438223630189896,
-0.13532517850399017,
-0.04527110606431961,
-0.012869495898485184,
0.014892865903675556,
-0.0015251357108354568,
0.014480562880635262,
-0.08388972282409668,
-0.05158208683133125,
-0.07699090987443924,
0.11613636463880539,
-0.1149277463555336,
0.011509082280099392,
0.08218037337064743,
-0.09468217939138412,
0.06472372263669968,
0.005267569329589605,
-0.013114025816321373,
-0.022969376295804977,
-0.10673835128545761,
0.002617563121020794,
0.016400013118982315,
0.05115996673703194,
-0.009346868842840195,
-0.2096521556377411,
0.001604026765562594,
-0.032098591327667236,
-0.023204460740089417,
-0.008197269402444363,
0.05617212504148483,
-0.10006187111139297,
0.039044152945280075,
-0.03497729077935219,
-0.010709146969020367,
-0.06132953613996506,
0.04699360206723213,
0.0902741551399231,
-0.011196027509868145,
0.1164148598909378,
-0.07110510766506195,
0.03928481414914131,
-0.12813276052474976,
-0.015591824427247047,
0.011447404511272907,
-0.009507409296929836,
-0.016796672716736794,
-0.027212250977754593,
0.07288716733455658,
0.0032548271119594574,
0.1988394409418106,
-0.013773995451629162,
-0.002083804225549102,
0.0628548413515091,
-0.04529924318194389,
0.009568965993821621,
0.055685315281152725,
0.0851120874285698,
0.036357104778289795,
0.010919857770204544,
-0.0021039661951363087,
0.008643225766718388,
0.032468315213918686,
-0.024744514375925064,
0.17054559290409088,
0.22423937916755676,
0.096621073782444,
0.07939367741346359,
-0.005889409687370062,
-0.08570461720228195,
-0.11064664274454117,
0.014906369149684906,
-0.03519834578037262,
0.10569903254508972,
-0.05398824065923691,
0.16924040019512177,
0.12917114794254303,
-0.20104745030403137,
0.0969223827123642,
-0.0578555092215538,
-0.07134262472391129,
-0.0931800827383995,
-0.12385629117488861,
-0.053803447633981705,
-0.08691299706697464,
-0.012969856150448322,
-0.10107105225324631,
0.04736572131514549,
0.10103697329759598,
0.0526166707277298,
0.020971208810806274,
0.07126601785421371,
-0.09029724448919296,
-0.04115350916981697,
0.08576466143131256,
0.015176651068031788,
0.017098434269428253,
-0.01063073705881834,
-0.0348738469183445,
0.020737288519740105,
-0.009905347600579262,
0.06281638890504837,
0.04020295664668083,
-0.0033528876956552267,
0.005116647109389305,
-0.013946626335382462,
-0.07378332316875458,
0.012735597789287567,
-0.018359845504164696,
-0.008621776476502419,
0.09159383177757263,
0.06807486712932587,
-0.031291186809539795,
-0.013691574335098267,
0.16084131598472595,
-0.06905263662338257,
-0.12826403975486755,
-0.1472092568874359,
0.15530240535736084,
-0.01906704343855381,
0.03201233223080635,
0.027188586071133614,
-0.10794119536876678,
-0.05520586296916008,
0.1703004390001297,
0.15470431745052338,
-0.04264608398079872,
-0.028440486639738083,
-0.0016253519570454955,
0.0001419595064362511,
-0.009116188623011112,
0.10830915719270706,
0.08484018594026566,
0.16401048004627228,
-0.06750702112913132,
0.06671708077192307,
-0.03930913284420967,
-0.08589375764131546,
-0.14055690169334412,
0.12360119819641113,
0.02317807450890541,
-0.02124103717505932,
-0.009177597239613533,
0.07440075278282166,
-0.05524761229753494,
-0.20307862758636475,
0.0023883080575615168,
-0.05916095897555351,
-0.13413166999816895,
-0.030390415340662003,
-0.018373854458332062,
0.06408713012933731,
0.031031230464577675,
0.02239677868783474,
-0.010363765992224216,
0.09752770513296127,
0.04507240653038025,
-0.06315803527832031,
-0.05067811533808708,
0.07650041580200195,
-0.11669240891933441,
0.18831339478492737,
0.02828608639538288,
0.07769627869129181,
0.10056424140930176,
-0.03369125723838806,
-0.13458512723445892,
0.0280073881149292,
0.07088328897953033,
-0.03658568859100342,
0.06408843398094177,
0.15831056237220764,
-0.004963811486959457,
0.03828826919198036,
0.05595630034804344,
-0.05756773427128792,
0.02398211695253849,
-0.015218839980661869,
-0.0461464561522007,
-0.07828172296285629,
0.025624344125390053,
-0.09240563213825226,
0.13907141983509064,
0.1820291429758072,
-0.04610820487141609,
0.018379660323262215,
-0.04060190171003342,
0.09259152412414551,
-0.023586316034197807,
0.03370651230216026,
-0.02645895630121231,
-0.17059341073036194,
0.0034933388233184814,
-0.05427400767803192,
0.0834958404302597,
-0.22588257491588593,
-0.07814758270978928,
-0.026479797437787056,
0.019453229382634163,
-0.06242822855710983,
0.12932223081588745,
0.11533161252737045,
0.02931877039372921,
-0.04567695036530495,
-0.1168171837925911,
-0.0017815687460824847,
0.13737379014492035,
-0.1282258927822113,
-0.03176892548799515
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mbart-sahit5rans
This model is a fine-tuned version of [google/mt5-base](https://huggingface.co/google/mt5-base) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 8.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"language": ["sa", "hi"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "google/mt5-base", "model-index": [{"name": "mbart-sahit5rans", "results": []}]} | text2text-generation | balaramas/mbart-sahit5rans | [
"transformers",
"safetensors",
"mt5",
"text2text-generation",
"generated_from_trainer",
"sa",
"hi",
"base_model:google/mt5-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T08:16:17+00:00 | [] | [
"sa",
"hi"
] | TAGS
#transformers #safetensors #mt5 #text2text-generation #generated_from_trainer #sa #hi #base_model-google/mt5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# mbart-sahit5rans
This model is a fine-tuned version of google/mt5-base on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 8.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# mbart-sahit5rans\n\nThis model is a fine-tuned version of google/mt5-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 8.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #mt5 #text2text-generation #generated_from_trainer #sa #hi #base_model-google/mt5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# mbart-sahit5rans\n\nThis model is a fine-tuned version of google/mt5-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 8.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
80,
33,
6,
12,
8,
3,
91,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #mt5 #text2text-generation #generated_from_trainer #sa #hi #base_model-google/mt5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# mbart-sahit5rans\n\nThis model is a fine-tuned version of google/mt5-base on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 8.0### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.08668193221092224,
0.13919423520565033,
-0.0022630987223237753,
0.10175016522407532,
0.14703667163848877,
-0.0025639974046498537,
0.11741480231285095,
0.13927003741264343,
-0.08263154327869415,
0.06454331427812576,
0.0818173885345459,
0.042533207684755325,
0.027672404423356056,
0.15330703556537628,
-0.03606211394071579,
-0.22363702952861786,
-0.005198813043534756,
-0.047055017203092575,
-0.04348801448941231,
0.10597043484449387,
0.09680936485528946,
-0.07858378440141678,
0.10538745671510696,
-0.0274004228413105,
-0.11950857192277908,
0.04912777617573738,
-0.02167200669646263,
-0.053856924176216125,
0.08570710569620132,
0.017029399052262306,
0.04694763198494911,
0.014327321201562881,
0.12523753941059113,
-0.23276028037071228,
-0.00007830728281987831,
0.07879806309938431,
0.022637823596596718,
0.07034014910459518,
0.050677426159381866,
-0.03976546600461006,
0.07677161693572998,
-0.14965498447418213,
0.08248833566904068,
0.046498946845531464,
-0.07344573736190796,
-0.11365571618080139,
-0.09043048322200775,
0.0675080195069313,
0.09219495952129364,
0.11986176669597626,
0.011067144572734833,
0.12171588093042374,
-0.07598225027322769,
0.08283083885908127,
0.22479571402072906,
-0.23486541211605072,
-0.046006955206394196,
0.07107739895582199,
0.04907951503992081,
0.0889902114868164,
-0.10622787475585938,
0.012134027667343616,
0.05780688673257828,
0.023782119154930115,
0.09021047502756119,
-0.018568433821201324,
-0.10782742500305176,
0.012430510483682156,
-0.11914221942424774,
-0.017346501350402832,
0.2030106484889984,
0.0448431670665741,
-0.040239106863737106,
-0.09444919228553772,
-0.06314139813184738,
-0.08743437379598618,
-0.010758606716990471,
-0.016124142333865166,
0.05554406717419624,
-0.0317230150103569,
-0.046392928808927536,
-0.07056865096092224,
-0.0686875656247139,
-0.07664786279201508,
0.014909750781953335,
0.07365617156028748,
0.03492224961519241,
0.0067801158875226974,
-0.03358222916722298,
0.10391833633184433,
-0.0180716123431921,
-0.12325979024171829,
-0.03509595990180969,
0.007491816766560078,
-0.045282550156116486,
-0.05144056677818298,
-0.016623958945274353,
-0.020558763295412064,
0.017711084336042404,
0.14838464558124542,
-0.07445141673088074,
0.0728643536567688,
-0.003300108714029193,
0.007252709940075874,
-0.032027602195739746,
0.11868437379598618,
-0.03605031222105026,
-0.07024391740560532,
0.04829977825284004,
0.09890342503786087,
0.04142843186855316,
0.0031199194490909576,
-0.08844935148954391,
-0.024978043511509895,
0.09286386519670486,
0.09546889364719391,
-0.012871036306023598,
0.05646246671676636,
-0.02372615970671177,
-0.00661853002384305,
0.03523564711213112,
-0.14556632936000824,
0.04404144734144211,
-0.0077897063456475735,
-0.0943717360496521,
-0.07756394147872925,
0.07837557792663574,
-0.006486881989985704,
-0.02855147235095501,
0.04758702963590622,
-0.06947492808103561,
0.005678894463926554,
-0.07389332354068756,
-0.04326251894235611,
0.035914305597543716,
-0.03615964949131012,
-0.0056765261106193066,
-0.08520030230283737,
-0.26366856694221497,
-0.04416993632912636,
0.020881567150354385,
-0.05622425675392151,
-0.04516996815800667,
-0.062344811856746674,
-0.07957825809717178,
-0.0006785767618566751,
-0.016012681648135185,
0.10628234595060349,
-0.057706404477357864,
0.06500769406557083,
0.022911962121725082,
0.02548985369503498,
0.0375954769551754,
0.044489841908216476,
-0.10859815031290054,
0.03772749751806259,
-0.15449322760105133,
0.0899936780333519,
-0.09456941485404968,
0.07105612754821777,
-0.11862345039844513,
-0.07104531675577164,
-0.018590226769447327,
-0.022386785596609116,
0.043621137738227844,
0.18055129051208496,
-0.14288201928138733,
-0.027808016166090965,
0.1275898665189743,
-0.07440469413995743,
-0.09934502840042114,
0.08832257241010666,
-0.029406065121293068,
0.05199655890464783,
0.08273322135210037,
0.14799870550632477,
0.0979413315653801,
-0.13670966029167175,
-0.02694348432123661,
-0.0019029079703614116,
0.04509694501757622,
0.004442022182047367,
0.054679617285728455,
-0.017439553514122963,
0.022686723619699478,
0.021453669294714928,
-0.06309118121862411,
0.008300730027258396,
-0.07483246922492981,
-0.08264791965484619,
-0.06774788349866867,
-0.09316031634807587,
0.013981659896671772,
0.03176772594451904,
0.04098737612366676,
-0.07747136801481247,
-0.10427550971508026,
0.09052729606628418,
0.1191045343875885,
-0.04903770610690117,
0.015201739966869354,
-0.05430839583277702,
0.03491911664605141,
-0.04340970143675804,
-0.008409164845943451,
-0.1842735856771469,
-0.13706204295158386,
0.041668396443128586,
-0.09039090573787689,
0.059271302074193954,
-0.05284483730792999,
0.07313620299100876,
0.07083571702241898,
-0.06951281428337097,
-0.037153054028749466,
-0.04738900437951088,
0.002271058736369014,
-0.10009024292230606,
-0.19149044156074524,
-0.025566885247826576,
-0.022084474563598633,
0.13894061744213104,
-0.24481970071792603,
0.039350833743810654,
-0.014986587688326836,
0.12939780950546265,
0.025638842955231667,
-0.04535214602947235,
0.028622908517718315,
0.013422410935163498,
-0.015924131497740746,
-0.10288105905056,
0.02596205472946167,
-0.011703308671712875,
-0.0693613588809967,
-0.019116580486297607,
-0.13500374555587769,
0.07736251503229141,
0.07662547379732132,
0.04718564450740814,
-0.0892241895198822,
-0.00008223999611800537,
-0.051471345126628876,
-0.053271327167749405,
-0.10385613888502121,
-0.012052151374518871,
0.1759520173072815,
0.0025022272020578384,
0.13023853302001953,
-0.06301576644182205,
-0.06335029006004333,
-0.007803929969668388,
-0.00839428324252367,
-0.021712200716137886,
0.09185021370649338,
0.05814098194241524,
-0.11437982320785522,
0.11422067135572433,
0.09537721425294876,
-0.033761847764253616,
0.1397642344236374,
-0.0420963391661644,
-0.07368634641170502,
-0.015185768716037273,
0.029113929718732834,
-0.02151874639093876,
0.08720274269580841,
-0.08486615866422653,
-0.009602280333638191,
0.008635289967060089,
0.02283729985356331,
0.043538957834243774,
-0.15295688807964325,
-0.0050734542310237885,
0.01625361293554306,
-0.06972070038318634,
-0.010435049422085285,
-0.020610902458429337,
0.007301389705389738,
0.0781540498137474,
0.014376485720276833,
-0.014084083028137684,
0.038162268698215485,
-0.007605820428580046,
-0.10299068689346313,
0.1820725053548813,
-0.10803992301225662,
-0.14641278982162476,
-0.13741876184940338,
0.03792473301291466,
-0.07402805238962173,
-0.017092229798436165,
0.025834251195192337,
-0.08016547560691833,
-0.05849751457571983,
-0.11998941749334335,
0.015946928411722183,
-0.018045974895358086,
-0.0267548318952322,
0.010662770830094814,
0.03852225840091705,
0.09218178689479828,
-0.12364615499973297,
0.007816175930202007,
0.01198615599423647,
-0.11298840492963791,
-0.029557164758443832,
0.028364649042487144,
0.10720816999673843,
0.10206978768110275,
-0.006600772961974144,
0.020185070112347603,
-0.023226166144013405,
0.20219969749450684,
-0.06520860642194748,
0.010459918528795242,
0.12731409072875977,
0.04466203227639198,
0.04660162702202797,
0.14059200882911682,
0.019530434161424637,
-0.10194941610097885,
0.05077391117811203,
0.0717272087931633,
-0.013074716553092003,
-0.2649899423122406,
-0.04326875880360603,
-0.014884646981954575,
-0.03352317214012146,
0.0823308527469635,
0.08351896703243256,
0.04100216180086136,
0.029758039861917496,
-0.023277457803487778,
0.04671217128634453,
0.016515163704752922,
0.07996241003274918,
0.08990888297557831,
0.025657624006271362,
0.07974948734045029,
-0.04191126674413681,
-0.013693234883248806,
0.06063506379723549,
0.010657642036676407,
0.26640549302101135,
-0.013138272799551487,
0.12254297733306885,
0.026049889624118805,
0.12155786901712418,
-0.017438417300581932,
0.01895090378820896,
0.049498800188302994,
0.014012741856276989,
0.013962075114250183,
-0.07534012943506241,
-0.01553257368505001,
0.05365331843495369,
-0.05250822752714157,
0.05564812198281288,
-0.08089131861925125,
0.04023958742618561,
0.037739839404821396,
0.26555341482162476,
0.01830524578690529,
-0.30016857385635376,
-0.07380802929401398,
0.03297719359397888,
-0.04564674198627472,
-0.03799629956483841,
0.030452623963356018,
0.10674185305833817,
-0.12830647826194763,
0.06944877654314041,
-0.04221165180206299,
0.0980825126171112,
-0.04531456157565117,
-0.0027946620248258114,
0.04030454903841019,
0.10563670098781586,
-0.015019255690276623,
0.09340426325798035,
-0.22998397052288055,
0.21317622065544128,
0.030455894768238068,
0.11541813611984253,
-0.047701794654130936,
0.02936944179236889,
0.035707443952560425,
0.13693423569202423,
0.12975358963012695,
-0.01218983344733715,
-0.0690760388970375,
-0.14115455746650696,
-0.10784191638231277,
0.027828248217701912,
0.10425884276628494,
-0.035566218197345734,
0.08234395831823349,
-0.05218442529439926,
-0.023392369970679283,
0.046934403479099274,
-0.08555863052606583,
-0.20247378945350647,
-0.098301962018013,
0.0002406001149211079,
0.013972871005535126,
0.006871173158288002,
-0.08837559819221497,
-0.09076710045337677,
-0.013848342932760715,
0.15505556762218475,
0.01912604831159115,
-0.05251244083046913,
-0.13954192399978638,
0.05708949640393257,
0.10809725522994995,
-0.07668772339820862,
0.01607729308307171,
0.01891854591667652,
0.1088099554181099,
0.04893569275736809,
-0.08895139396190643,
0.06939636915922165,
-0.07570512592792511,
-0.1722322702407837,
-0.045537445694208145,
0.1356486827135086,
0.043058786541223526,
0.04177705571055412,
-0.013518801890313625,
0.025891367346048355,
0.009389081969857216,
-0.08646070212125778,
0.006135033909231424,
0.10467001795768738,
0.0743335634469986,
0.0863257646560669,
-0.09600292891263962,
-0.000309887807816267,
-0.042416319251060486,
-0.041165485978126526,
0.10215232521295547,
0.20122624933719635,
-0.0699307918548584,
0.08642469346523285,
0.09310215711593628,
-0.09746061265468597,
-0.2082020342350006,
0.06993135064840317,
0.054516684263944626,
-0.0026571510825306177,
0.05333336442708969,
-0.17763936519622803,
0.10190458595752716,
0.09791573137044907,
-0.036579959094524384,
0.0846334770321846,
-0.32329079508781433,
-0.13848288357257843,
0.05804576724767685,
0.09503673762083054,
-0.011376705020666122,
-0.159726083278656,
-0.06182454898953438,
-0.0512322299182415,
-0.11370164901018143,
0.06476938724517822,
-0.10737974941730499,
0.10704253613948822,
-0.0066590504720807076,
0.08984550833702087,
0.013530063442885876,
-0.041497986763715744,
0.12312100827693939,
0.03163575753569603,
0.0959121435880661,
-0.06738436222076416,
0.01868302933871746,
0.08822844922542572,
-0.07753734290599823,
0.09720630943775177,
-0.05746576189994812,
0.07991614937782288,
-0.11411507427692413,
-0.02401777356863022,
-0.06457875669002533,
0.09032806754112244,
-0.053282737731933594,
-0.042113590985536575,
-0.05089308321475983,
0.04191743582487106,
0.05168354883790016,
-0.022245584055781364,
0.10055551677942276,
0.0031968113034963608,
0.08183155208826065,
0.1314237117767334,
0.11412002891302109,
-0.011630071327090263,
-0.08133222907781601,
0.005339323543012142,
-0.034569647163152695,
0.0398029126226902,
-0.18184947967529297,
0.02962464839220047,
0.1067684069275856,
0.04026186093688011,
0.14042969048023224,
0.025596849620342255,
-0.05791161581873894,
-0.005705187097191811,
0.03958229720592499,
-0.11621134728193283,
-0.16030320525169373,
-0.03618355467915535,
-0.07167771458625793,
-0.11674391478300095,
0.035577915608882904,
0.09996654838323593,
-0.09727409482002258,
-0.004770670086145401,
-0.027310779318213463,
0.05071672052145004,
-0.021908853203058243,
0.18490375578403473,
0.05222000554203987,
0.048087235540151596,
-0.0834631621837616,
0.1749085783958435,
0.07833509892225266,
-0.045933205634355545,
0.04594877362251282,
0.10970745980739594,
-0.09267239272594452,
-0.03988926112651825,
0.057030484080314636,
0.18515457212924957,
-0.02354910597205162,
-0.05034654960036278,
-0.10115130990743637,
-0.0724959596991539,
0.04445665702223778,
0.12052590399980545,
0.0380173921585083,
-0.00853396113961935,
-0.003602837910875678,
0.021052222698926926,
-0.12793809175491333,
0.11600041389465332,
0.04245305061340332,
0.06302843987941742,
-0.16893678903579712,
0.0728297084569931,
0.01447147224098444,
0.03810552880167961,
-0.02494848147034645,
0.034554142504930496,
-0.08367010205984116,
-0.01449454203248024,
-0.12427125126123428,
-0.008349468931555748,
-0.018542053177952766,
0.006257010158151388,
-0.022227125242352486,
-0.06125869229435921,
-0.052383020520210266,
0.06644921004772186,
-0.06091923639178276,
-0.07268138974905014,
0.003895208239555359,
0.0735509917140007,
-0.14950619637966156,
-0.023989999666810036,
0.034027349203825,
-0.08642667531967163,
0.07393255829811096,
0.06798375397920609,
0.011211122386157513,
0.03141331300139427,
-0.13069064915180206,
-0.0018692872254177928,
0.03549399971961975,
0.015755299478769302,
0.03952622413635254,
-0.10612469911575317,
-0.014692013151943684,
-0.017474552616477013,
0.04101371392607689,
0.017492081969976425,
0.04510492458939552,
-0.12310678511857986,
-0.018128281459212303,
-0.06429162621498108,
-0.03370673954486847,
-0.05793321505188942,
0.04337085410952568,
0.10954933613538742,
0.011607336811721325,
0.1592625081539154,
-0.09654480218887329,
0.037960201501846313,
-0.20429037511348724,
-0.02228419855237007,
-0.0002893893397413194,
-0.05295589193701744,
-0.10255097597837448,
-0.019511301070451736,
0.08152494579553604,
-0.05564146488904953,
0.1381746530532837,
-0.00090616854140535,
0.08328157663345337,
0.047841183841228485,
-0.011517313309013844,
-0.002392168389633298,
0.01107020303606987,
0.13629165291786194,
0.04808581992983818,
-0.00560367526486516,
0.09812265634536743,
-0.00393125182017684,
0.08317792415618896,
-0.01116363424807787,
0.15841734409332275,
0.14791442453861237,
-0.04616985097527504,
0.08160349726676941,
0.0819077342748642,
-0.11747981607913971,
-0.13191670179367065,
0.08577361702919006,
-0.019848981872200966,
0.11074326932430267,
-0.07531822472810745,
0.12524186074733734,
0.1276322454214096,
-0.14374974370002747,
0.047944504767656326,
-0.06833098083734512,
-0.09770427644252777,
-0.12521260976791382,
-0.04496219754219055,
-0.09483716636896133,
-0.14031794667243958,
0.02439609169960022,
-0.10799166560173035,
0.04037361592054367,
0.06683511286973953,
-0.003923141397535801,
-0.0019916361197829247,
0.17982259392738342,
-0.022724376991391182,
0.0034936792217195034,
0.047022510319948196,
-0.011593600735068321,
-0.015496579930186272,
-0.06131517514586449,
-0.0527745820581913,
0.052869923412799835,
-0.006656516809016466,
0.047015924006700516,
-0.01917392760515213,
-0.009279096499085426,
0.041973814368247986,
-0.02501806430518627,
-0.06934068351984024,
0.014612958766520023,
0.034160174429416656,
0.01885361038148403,
0.044810567051172256,
0.05313614755868912,
-0.01877191849052906,
-0.012670404277741909,
0.30567243695259094,
-0.08154075592756271,
-0.07932385057210922,
-0.10922635346651077,
0.17637696862220764,
0.03907303139567375,
-0.00636523962020874,
0.039804328233003616,
-0.127531036734581,
0.008413576520979404,
0.17080245912075043,
0.15915852785110474,
-0.037978705018758774,
-0.012302685528993607,
-0.03222483769059181,
-0.018399149179458618,
-0.03158468008041382,
0.08859238773584366,
0.10304030030965805,
0.01683792471885681,
-0.04721663519740105,
-0.00041289004730060697,
0.014948631636798382,
-0.03879103064537048,
-0.10186303406953812,
0.09497697651386261,
-0.01908983662724495,
0.021817393600940704,
-0.023925496265292168,
0.0841650515794754,
-0.019060837104916573,
-0.17638172209262848,
0.01205387618392706,
-0.1653660386800766,
-0.14623373746871948,
-0.03443920612335205,
0.08971678465604782,
0.0010941714281216264,
0.03833942487835884,
0.0011567616602405906,
-0.00035581892007030547,
0.11768122017383575,
-0.02209409698843956,
-0.09713125228881836,
-0.11380311101675034,
0.05090881511569023,
-0.08262192457914352,
0.2510952353477478,
-0.0014921212568879128,
0.029302000999450684,
0.10231611877679825,
-0.025316426530480385,
-0.1592622548341751,
0.07297196239233017,
0.059672825038433075,
-0.054091449826955795,
0.04047612473368645,
0.15262170135974884,
-0.043771132826805115,
0.10931772738695145,
0.02236093394458294,
-0.07056485861539841,
-0.008635800331830978,
-0.0029575014486908913,
-0.031875357031822205,
-0.10462089627981186,
-0.015871841460466385,
-0.06192592903971672,
0.1500277817249298,
0.16566069424152374,
-0.04209352657198906,
0.024512173607945442,
-0.08293109387159348,
0.03154623135924339,
0.05523679777979851,
0.06942513585090637,
0.0027669165283441544,
-0.2121913582086563,
0.024142449721693993,
0.0405263788998127,
0.03959617018699646,
-0.23846735060214996,
-0.08099900931119919,
0.03866041079163551,
-0.04549175873398781,
-0.06729446351528168,
0.11073940992355347,
0.10179487615823746,
0.028756851330399513,
-0.03149314969778061,
-0.1598842442035675,
-0.02789974957704544,
0.16559843719005585,
-0.1687254160642624,
-0.05772874876856804
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# lmind_nq_train10000_eval6489_v1_qa_tyzhu_lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl
This model is a fine-tuned version of [tyzhu/lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl](https://huggingface.co/tyzhu/lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl) on the tyzhu/lmind_nq_train10000_eval6489_v1_qa dataset.
It achieves the following results on the evaluation set:
- Loss: 3.0531
- Accuracy: 0.5457
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant
- num_epochs: 20.0
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 2.109 | 1.0 | 625 | 2.0393 | 0.5570 |
| 1.5029 | 2.0 | 1250 | 2.0911 | 0.5576 |
| 1.0521 | 3.0 | 1875 | 2.2873 | 0.5567 |
| 0.8245 | 4.0 | 2500 | 2.4180 | 0.5545 |
| 0.7322 | 5.0 | 3125 | 2.4951 | 0.5484 |
| 0.6771 | 6.0 | 3750 | 2.5769 | 0.5539 |
| 0.6445 | 7.0 | 4375 | 2.6174 | 0.5541 |
| 0.614 | 8.0 | 5000 | 2.6924 | 0.5539 |
| 0.5966 | 9.0 | 5625 | 2.7009 | 0.5529 |
| 0.58 | 10.0 | 6250 | 2.7747 | 0.5515 |
| 0.5657 | 11.0 | 6875 | 2.7767 | 0.5507 |
| 0.5585 | 12.0 | 7500 | 2.8466 | 0.5503 |
| 0.5432 | 13.0 | 8125 | 2.8841 | 0.5502 |
| 0.5284 | 14.0 | 8750 | 2.9405 | 0.5498 |
| 0.5242 | 15.0 | 9375 | 2.8969 | 0.5491 |
| 0.511 | 16.0 | 10000 | 2.9666 | 0.5480 |
| 0.5051 | 17.0 | 10625 | 2.9805 | 0.5487 |
| 0.4947 | 18.0 | 11250 | 2.9896 | 0.5476 |
| 0.4832 | 19.0 | 11875 | 2.9937 | 0.5473 |
| 0.4803 | 20.0 | 12500 | 3.0531 | 0.5457 |
### Framework versions
- Transformers 4.34.0
- Pytorch 2.1.0+cu121
- Datasets 2.14.5
- Tokenizers 0.14.1
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["tyzhu/lmind_nq_train10000_eval6489_v1_qa"], "metrics": ["accuracy"], "base_model": "tyzhu/lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl", "model-index": [{"name": "lmind_nq_train10000_eval6489_v1_qa_tyzhu_lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl", "results": [{"task": {"type": "text-generation", "name": "Causal Language Modeling"}, "dataset": {"name": "tyzhu/lmind_nq_train10000_eval6489_v1_qa", "type": "tyzhu/lmind_nq_train10000_eval6489_v1_qa"}, "metrics": [{"type": "accuracy", "value": 0.5456769049002058, "name": "Accuracy"}]}]}]} | text-generation | tyzhu/lmind_nq_train10000_eval6489_v1_qa_tyzhu_lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"generated_from_trainer",
"dataset:tyzhu/lmind_nq_train10000_eval6489_v1_qa",
"base_model:tyzhu/lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl",
"license:mit",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-06T08:23:13+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #generated_from_trainer #dataset-tyzhu/lmind_nq_train10000_eval6489_v1_qa #base_model-tyzhu/lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl #license-mit #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| lmind\_nq\_train10000\_eval6489\_v1\_qa\_tyzhu\_lmind\_nq\_train10000\_eval6489\_v1\_docidx\_gpt2-xl
====================================================================================================
This model is a fine-tuned version of tyzhu/lmind\_nq\_train10000\_eval6489\_v1\_docidx\_gpt2-xl on the tyzhu/lmind\_nq\_train10000\_eval6489\_v1\_qa dataset.
It achieves the following results on the evaluation set:
* Loss: 3.0531
* Accuracy: 0.5457
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: constant
* num\_epochs: 20.0
### Training results
### Framework versions
* Transformers 4.34.0
* Pytorch 2.1.0+cu121
* Datasets 2.14.5
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* num\\_epochs: 20.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.34.0\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.5\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #generated_from_trainer #dataset-tyzhu/lmind_nq_train10000_eval6489_v1_qa #base_model-tyzhu/lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl #license-mit #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* num\\_epochs: 20.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.34.0\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.5\n* Tokenizers 0.14.1"
] | [
125,
99,
4,
33
] | [
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #generated_from_trainer #dataset-tyzhu/lmind_nq_train10000_eval6489_v1_qa #base_model-tyzhu/lmind_nq_train10000_eval6489_v1_docidx_gpt2-xl #license-mit #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\n* num\\_epochs: 20.0### Training results### Framework versions\n\n\n* Transformers 4.34.0\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.5\n* Tokenizers 0.14.1"
] | [
-0.08655204623937607,
0.06521283090114594,
-0.0037300502881407738,
0.08564288914203644,
0.14113466441631317,
0.04371725022792816,
0.14187544584274292,
0.13558922708034515,
-0.133244588971138,
0.060869261622428894,
0.12653748691082,
0.1292819380760193,
0.03632030636072159,
0.07814040780067444,
-0.01648619957268238,
-0.2494983673095703,
-0.005044727586209774,
0.032856494188308716,
-0.12911684811115265,
0.12283738702535629,
0.06842870265245438,
-0.13019990921020508,
0.07832450419664383,
0.009296038188040257,
-0.18210813403129578,
0.010752578265964985,
-0.025511687621474266,
-0.04053991660475731,
0.1254831701517105,
0.033893000334501266,
0.09562572091817856,
0.020053880289196968,
0.10367303341627121,
-0.2333906590938568,
0.00304687162861228,
0.07793942093849182,
-0.008679819293320179,
0.05718491971492767,
0.04121841490268707,
-0.02194800041615963,
0.1827802211046219,
-0.07136683166027069,
0.06328979879617691,
0.016499612480401993,
-0.13454300165176392,
-0.22438326478004456,
-0.06798259913921356,
0.05158409848809242,
0.0656195804476738,
0.11503563076257706,
-0.024247899651527405,
0.10484487563371658,
-0.08737903088331223,
0.09886761754751205,
0.2498633712530136,
-0.24350272119045258,
-0.08581720292568207,
0.08262861520051956,
0.01097782514989376,
0.06561695039272308,
-0.11120784282684326,
-0.018145151436328888,
0.05018961802124977,
0.033446624875068665,
0.10184076428413391,
0.00013056657917331904,
-0.03017316199839115,
0.04118159040808678,
-0.13262057304382324,
-0.034393250942230225,
0.11972980201244354,
0.047003235667943954,
-0.0031459247693419456,
-0.036446232348680496,
-0.047260746359825134,
-0.20700299739837646,
-0.0309569351375103,
-0.03136083111166954,
0.05041123926639557,
-0.030839618295431137,
-0.09435173869132996,
0.017131948843598366,
-0.0750059112906456,
-0.08304229378700256,
-0.0005347231053747237,
0.10330784320831299,
0.060867149382829666,
-0.01880965754389763,
-0.03138290345668793,
0.1259375512599945,
-0.013235603459179401,
-0.14310085773468018,
-0.010475494898855686,
0.03820164129137993,
-0.020781217142939568,
-0.04082471504807472,
-0.05509861931204796,
-0.02439771220088005,
0.01562616042792797,
0.12601475417613983,
-0.05508881062269211,
0.0742841437458992,
0.03519102558493614,
0.010675624944269657,
-0.10263775289058685,
0.1641274392604828,
-0.057473279535770416,
-0.039388153702020645,
-0.02589656598865986,
0.08687475323677063,
-0.0265929214656353,
-0.0101885125041008,
-0.08534512668848038,
0.030208170413970947,
0.11060867458581924,
0.0028522571083158255,
-0.07211409509181976,
0.07787951081991196,
-0.040413156151771545,
-0.01408638246357441,
-0.06640169769525528,
-0.07257213443517685,
0.032782990485429764,
0.019035568460822105,
-0.08504407107830048,
-0.0088212750852108,
0.014238866977393627,
0.014126082882285118,
-0.006276503670960665,
0.12296849489212036,
-0.09841190278530121,
0.017301952466368675,
-0.11095248907804489,
-0.10941507667303085,
-0.0022817954886704683,
-0.04907578229904175,
0.017677105963230133,
-0.0661396011710167,
-0.17207999527454376,
-0.04866752400994301,
0.055239900946617126,
-0.03539165109395981,
-0.030464470386505127,
-0.09182330220937729,
-0.11381959170103073,
0.03152003139257431,
0.01545706856995821,
0.12527289986610413,
-0.05991015210747719,
0.10089676082134247,
0.03598076105117798,
0.049945347011089325,
0.04847311973571777,
0.04855186119675636,
-0.06836161762475967,
0.018899301066994667,
-0.1300407499074936,
0.06194464862346649,
-0.042511649429798126,
0.03529670089483261,
-0.08166613429784775,
-0.12108202278614044,
-0.028138186782598495,
0.005635345354676247,
0.08272532373666763,
0.09873169660568237,
-0.1880583018064499,
-0.09210891276597977,
0.20165245234966278,
-0.053225450217723846,
-0.12193556874990463,
0.09683854877948761,
-0.0579964704811573,
0.014336301013827324,
0.06647209078073502,
0.13386660814285278,
0.06523680686950684,
-0.08109410107135773,
-0.0089533356949687,
-0.05766880139708519,
0.09589312970638275,
0.05127408355474472,
0.04631080478429794,
0.009594560600817204,
0.055109065026044846,
0.012396988458931446,
-0.04769878461956978,
0.02035762183368206,
-0.12591049075126648,
-0.09026540815830231,
-0.03583799675107002,
-0.07714205980300903,
0.0030927439220249653,
0.06275203078985214,
0.06261961162090302,
-0.09173782914876938,
-0.10894559323787689,
0.08248958736658096,
0.0941016674041748,
-0.095378577709198,
0.01329796388745308,
-0.06239820644259453,
0.06246303394436836,
-0.009769548662006855,
0.00005396956839831546,
-0.13936322927474976,
-0.04925873875617981,
0.008650587871670723,
0.021656319499015808,
0.006577946245670319,
0.05350262671709061,
0.08601798862218857,
0.08653866499662399,
-0.05800522118806839,
-0.042667802423238754,
-0.10211693495512009,
-0.006934044416993856,
-0.10155095905065536,
-0.2331499606370926,
-0.01900380104780197,
-0.001317324466072023,
0.1615220010280609,
-0.2520699203014374,
0.04967562481760979,
0.008806667290627956,
0.0739387646317482,
0.01215712632983923,
-0.048151083290576935,
-0.0543544739484787,
0.07693717628717422,
-0.04913909733295441,
-0.06674904376268387,
0.05179158225655556,
-0.012803432531654835,
-0.08474673330783844,
-0.06905267387628555,
-0.14649349451065063,
0.16582569479942322,
0.12366870045661926,
-0.0764351487159729,
-0.10779298841953278,
0.0003246472915634513,
-0.066670261323452,
-0.03664400801062584,
-0.03229597583413124,
0.0009984043426811695,
0.20342089235782623,
-0.015762455761432648,
0.12722323834896088,
-0.07303954660892487,
-0.056866299360990524,
0.021155379712581635,
-0.008387558162212372,
0.03948583826422691,
0.14507140219211578,
0.1162794977426529,
-0.012014682404696941,
0.13945870101451874,
0.06022466719150543,
-0.06532593816518784,
0.1583162248134613,
-0.03358406946063042,
-0.06979147344827652,
-0.04381708800792694,
0.011324002407491207,
0.00826941803097725,
0.08299797773361206,
-0.1587819755077362,
-0.008832234889268875,
-0.007000131066888571,
0.034973058849573135,
0.023321613669395447,
-0.22268442809581757,
-0.05015191063284874,
0.0658431202173233,
-0.060707349330186844,
-0.01669257879257202,
-0.015412447042763233,
-0.006799120455980301,
0.09982117265462875,
-0.0018396021332591772,
-0.046243201941251755,
-0.01642456464469433,
-0.000825501570943743,
-0.0869092270731926,
0.21836888790130615,
-0.08915599435567856,
-0.0828622579574585,
-0.07622990012168884,
-0.03809488192200661,
-0.036767613142728806,
0.014019954018294811,
0.05806761980056763,
-0.1366426944732666,
-0.010533943772315979,
-0.06336868554353714,
0.0648493692278862,
-0.03094935417175293,
0.018116293475031853,
0.029687121510505676,
-0.0013147337595000863,
0.0432840995490551,
-0.12394987046718597,
-0.00941639207303524,
-0.04794066771864891,
-0.10774414986371994,
0.05832529813051224,
0.03548054397106171,
0.10907880961894989,
0.14170026779174805,
0.006969965994358063,
0.022119203582406044,
-0.015993809327483177,
0.2567061185836792,
-0.07925678044557571,
-0.022268695756793022,
0.0984354168176651,
0.02988172136247158,
0.04193198308348656,
0.08891643583774567,
0.07108698040246964,
-0.10927003622055054,
0.010030213743448257,
0.07141616195440292,
-0.03211234509944916,
-0.2101665586233139,
-0.004013689234852791,
-0.048127543181180954,
-0.041951462626457214,
0.08740822225809097,
0.0171285942196846,
0.0382116474211216,
0.06617892533540726,
0.007613875437527895,
0.09080490469932556,
-0.019800081849098206,
0.07173897325992584,
0.06022047996520996,
0.053676996380090714,
0.11780594289302826,
-0.010719958692789078,
-0.06330899149179459,
0.01994783990085125,
-0.03419383242726326,
0.24917802214622498,
-0.004684742074459791,
0.09694354236125946,
0.0793067067861557,
0.13910089433193207,
-0.018816472962498665,
0.0448460727930069,
-0.015530898235738277,
-0.052366726100444794,
-0.009285963140428066,
-0.05615188926458359,
-0.028626050800085068,
0.035399239510297775,
-0.037503648549318314,
0.06994111835956573,
-0.11596498638391495,
0.023937102407217026,
0.06495150923728943,
0.24356165528297424,
0.030593838542699814,
-0.3174920380115509,
-0.08153201639652252,
-0.006504008546471596,
-0.002413723384961486,
-0.012115808203816414,
0.01368142943829298,
0.088193379342556,
-0.1047905683517456,
0.03003225475549698,
-0.0511879026889801,
0.08406487852334976,
-0.04459284618496895,
0.03843425214290619,
0.058443568646907806,
0.09186075627803802,
-0.00757435942068696,
0.06869730353355408,
-0.3017425239086151,
0.26681751012802124,
-0.0016808899817988276,
0.0991034060716629,
-0.050390638411045074,
0.017647728323936462,
0.028982628136873245,
0.032095808535814285,
0.04093300551176071,
-0.004159846808761358,
-0.022599687799811363,
-0.20952416956424713,
-0.03223090618848801,
0.0394030436873436,
0.12105955183506012,
-0.0742265060544014,
0.13442179560661316,
-0.014302442781627178,
0.009394840337336063,
0.07192465662956238,
0.009330332279205322,
-0.09076255559921265,
-0.09629267454147339,
0.004414279013872147,
-0.024729173630475998,
0.013116307556629181,
-0.07003092020750046,
-0.10151845216751099,
-0.13308721780776978,
0.1524997502565384,
-0.04408492520451546,
-0.029456129297614098,
-0.10709136724472046,
0.10925140976905823,
0.08526665717363358,
-0.0799473226070404,
0.024758704006671906,
0.030120112001895905,
0.061360154300928116,
0.03858688473701477,
-0.0339149609208107,
0.09269563853740692,
-0.07063990086317062,
-0.20390012860298157,
-0.06595616787672043,
0.10168015956878662,
0.06458316743373871,
0.0706578865647316,
-0.024529440328478813,
0.027184581384062767,
-0.027003737166523933,
-0.10257323831319809,
0.05196760594844818,
0.027698643505573273,
0.0762164294719696,
0.012612858787178993,
-0.07464544475078583,
0.08654289692640305,
-0.057669300585985184,
-0.03897758200764656,
0.12172158062458038,
0.28019362688064575,
-0.07169567048549652,
0.0020732588600367308,
-0.002588113769888878,
-0.09076840430498123,
-0.13228502869606018,
0.06853824853897095,
0.09658128023147583,
0.011041996069252491,
0.008568234741687775,
-0.23934997618198395,
0.0961303785443306,
0.1156398355960846,
0.016322704032063484,
0.11465844511985779,
-0.37045395374298096,
-0.12934808433055878,
0.0625872015953064,
0.14278508722782135,
0.06074943020939827,
-0.16726043820381165,
-0.022638048976659775,
-0.004507908131927252,
-0.14956365525722504,
0.07769282162189484,
-0.07506915181875229,
0.12223722785711288,
-0.03711724653840065,
0.05826854705810547,
0.020518654957413673,
-0.05583481118083,
0.12878772616386414,
0.015974272042512894,
0.10593890398740768,
-0.05034780502319336,
-0.0050184293650090694,
0.050765376538038254,
-0.043242331594228745,
-0.012505094520747662,
-0.09940312802791595,
0.03687561675906181,
-0.09263485670089722,
-0.027562499046325684,
-0.08246475458145142,
0.0129062719643116,
-0.06325796991586685,
-0.07342147827148438,
-0.04469157010316849,
0.02554909698665142,
0.08687403798103333,
-0.018612714484333992,
0.09230439364910126,
-0.006589841563254595,
0.1820930540561676,
0.03354814648628235,
0.03726256638765335,
-0.01627584919333458,
-0.07081925868988037,
-0.005991256795823574,
0.02512458711862564,
0.05331333726644516,
-0.1470666527748108,
0.01233069971203804,
0.15699133276939392,
0.03147203102707863,
0.16579154133796692,
0.10035241395235062,
-0.050011616200208664,
0.013035954907536507,
0.06053332984447479,
-0.1238541230559349,
-0.10125810652971268,
-0.01251369807869196,
-0.08139345049858093,
-0.08887093514204025,
0.028084274381399155,
0.0930425301194191,
-0.07961087673902512,
0.0012727632420137525,
-0.009828543290495872,
0.014856366440653801,
-0.03906095772981644,
0.2191319465637207,
0.02198166586458683,
0.06296883523464203,
-0.09038865566253662,
0.0492851659655571,
0.01660960167646408,
-0.055174730718135834,
0.039783354848623276,
0.05202431231737137,
-0.06887225061655045,
-0.011244525201618671,
0.09844935685396194,
0.19771882891654968,
-0.0450136736035347,
-0.011884813196957111,
-0.1503286063671112,
-0.10960603505373001,
0.08316098153591156,
0.13740117847919464,
0.06922239810228348,
0.011466596275568008,
-0.040191616863012314,
0.0519791878759861,
-0.11316635459661484,
0.09069760143756866,
0.04763830453157425,
0.0666918233036995,
-0.11454657465219498,
0.1892816722393036,
0.017962642014026642,
-0.007585742510855198,
0.0025914940051734447,
0.020952288061380386,
-0.13409672677516937,
0.013768567703664303,
-0.10365539789199829,
-0.04668193683028221,
-0.013888690620660782,
-0.009847023524343967,
-0.006376955192536116,
-0.04093535989522934,
-0.05663054808974266,
0.02131286822259426,
-0.10870758444070816,
-0.04639323055744171,
0.010344147682189941,
0.0614490881562233,
-0.11335175484418869,
-0.010265875607728958,
0.025444883853197098,
-0.09643534570932388,
0.06744787842035294,
0.06535115092992783,
0.023063726723194122,
0.051485538482666016,
-0.15030218660831451,
0.04086781293153763,
0.04025670886039734,
0.0088032316416502,
0.04155260697007179,
-0.12353647500276566,
0.006436173804104328,
-0.04528271034359932,
0.03765084594488144,
0.019691042602062225,
0.030359040945768356,
-0.12122207880020142,
0.0023588768672198057,
-0.04052014648914337,
-0.07282458245754242,
-0.05430183932185173,
0.04387914389371872,
0.06075681000947952,
0.009967022575438023,
0.16446730494499207,
-0.07571824640035629,
0.050892919301986694,
-0.236899733543396,
-0.023128869011998177,
0.021549420431256294,
-0.07402494549751282,
-0.08178611099720001,
-0.058782465755939484,
0.07873363047838211,
-0.07364627718925476,
0.1050427258014679,
-0.02362477406859398,
0.008407137356698513,
0.013772845268249512,
-0.07265951484441757,
0.00026027727290056646,
0.01016988791525364,
0.207534521818161,
0.056419216096401215,
-0.0340174175798893,
0.045685701072216034,
0.05091177672147751,
0.0974821224808693,
0.10602367669343948,
0.19836483895778656,
0.12197059392929077,
-0.01921651139855385,
0.08239788562059402,
0.03759089484810829,
-0.10477861016988754,
-0.1384119689464569,
0.08525336533784866,
-0.0609828382730484,
0.11041931062936783,
-0.0280709620565176,
0.1929338127374649,
0.12761686742305756,
-0.17250318825244904,
0.050493888556957245,
-0.05028306320309639,
-0.0859718844294548,
-0.1072864904999733,
-0.028994211927056313,
-0.08719167113304138,
-0.16783079504966736,
0.03984059393405914,
-0.1392504870891571,
0.04330076649785042,
0.07812006026506424,
0.041374802589416504,
0.021641019731760025,
0.1584707498550415,
0.05652128532528877,
0.010925640352070332,
0.03605964034795761,
0.012874728068709373,
-0.03350114822387695,
-0.09909073263406754,
-0.07132411003112793,
0.013375939801335335,
-0.04845631867647171,
0.04890643432736397,
-0.04444151744246483,
-0.0635446161031723,
0.034288737922906876,
-0.005672454833984375,
-0.07786042243242264,
0.029895015060901642,
-0.008428416214883327,
0.061978381127119064,
0.036251965910196304,
0.0063288286328315735,
0.0008041512919589877,
-0.020471174269914627,
0.22239618003368378,
-0.06818602234125137,
-0.07674022763967514,
-0.10383245348930359,
0.2250441014766693,
0.06673457473516464,
0.00806169118732214,
0.03874335438013077,
-0.07274240255355835,
0.012925507500767708,
0.22167953848838806,
0.18357501924037933,
-0.07026547193527222,
-0.0075418478809297085,
0.020852113142609596,
-0.013731507584452629,
-0.019599098712205887,
0.08125535398721695,
0.12312235683202744,
0.04120206832885742,
-0.06681028753519058,
-0.030475961044430733,
-0.05427839607000351,
-0.021702120080590248,
-0.007334731053560972,
0.07597523927688599,
0.055092573165893555,
0.0028925840742886066,
-0.02446538768708706,
0.05247383192181587,
-0.05655098333954811,
-0.11836820095777512,
0.07334169000387192,
-0.2053394913673401,
-0.15728144347667694,
-0.016601094976067543,
0.05997119098901749,
0.004371268674731255,
0.08223778754472733,
-0.019873328506946564,
-0.04123426973819733,
0.07300551235675812,
-0.010784794576466084,
-0.06431111693382263,
-0.10759702324867249,
0.07583688199520111,
-0.0678749531507492,
0.18932783603668213,
-0.05602315440773964,
0.035712458193302155,
0.13771556317806244,
0.05249473825097084,
-0.04930330812931061,
0.06731563061475754,
0.06461447477340698,
-0.10504764318466187,
-0.007110038306564093,
0.10449996590614319,
-0.045993246138095856,
0.0778547152876854,
0.062123432755470276,
-0.12202376872301102,
0.03359321132302284,
-0.013140964321792126,
-0.05139145255088806,
-0.048840802162885666,
-0.026402849704027176,
-0.050065528601408005,
0.12684553861618042,
0.25399982929229736,
-0.025971824303269386,
0.011478371918201447,
-0.06972428411245346,
0.037849076092243195,
0.049580398947000504,
0.0887456014752388,
-0.05602886900305748,
-0.26754650473594666,
0.029076064005494118,
0.06970757991075516,
-0.007316689006984234,
-0.22531434893608093,
-0.11097685247659683,
0.05530712008476257,
-0.06952306628227234,
-0.07597093284130096,
0.09758171439170837,
0.10166668891906738,
0.06914190202951431,
-0.06293737143278122,
-0.08549581468105316,
-0.07097295671701431,
0.18520957231521606,
-0.14050954580307007,
-0.10416565835475922
] |
null | null | peft | ## Training procedure
### Framework versions
- PEFT 0.4.0
| {"library_name": "peft"} | null | lourenswal/bloom_prompt_tuning_1707207951.865479 | [
"peft",
"safetensors",
"region:us"
] | 2024-02-06T08:25:18+00:00 | [] | [] | TAGS
#peft #safetensors #region-us
| ## Training procedure
### Framework versions
- PEFT 0.4.0
| [
"## Training procedure",
"### Framework versions\n\n\n- PEFT 0.4.0"
] | [
"TAGS\n#peft #safetensors #region-us \n",
"## Training procedure",
"### Framework versions\n\n\n- PEFT 0.4.0"
] | [
14,
3,
11
] | [
"passage: TAGS\n#peft #safetensors #region-us \n## Training procedure### Framework versions\n\n\n- PEFT 0.4.0"
] | [
-0.11905625462532043,
0.0345856249332428,
-0.005941327195614576,
0.021338168531656265,
0.1318943053483963,
-0.04500391334295273,
0.11790177971124649,
0.044265083968639374,
0.02826196327805519,
0.04287289083003998,
0.14522042870521545,
0.012527059763669968,
-0.002473803237080574,
0.23268604278564453,
-0.02491838112473488,
-0.24689897894859314,
0.0963108167052269,
-0.018648667261004448,
0.010325984098017216,
0.1300942450761795,
0.06127513200044632,
-0.05820246413350105,
-0.00450087571516633,
-0.08924039453268051,
-0.12434796243906021,
0.030397437512874603,
0.0022219268139451742,
-0.05412381514906883,
0.15582223236560822,
-0.04505776986479759,
0.22642453014850616,
-0.026251045987010002,
0.03009611926972866,
-0.24157112836837769,
0.02298162505030632,
0.09180472791194916,
-0.014651010744273663,
0.05359065905213356,
0.08202625811100006,
0.00805692095309496,
0.013744879513978958,
0.006256286054849625,
0.06165527552366257,
0.04658261314034462,
-0.18916983902454376,
-0.16644287109375,
-0.10306265950202942,
0.08746249973773956,
0.19676101207733154,
0.09857923537492752,
0.038025811314582825,
0.1590358465909958,
-0.16483746469020844,
0.03241099417209625,
0.12431377917528152,
-0.32048213481903076,
-0.09287982434034348,
0.12529990077018738,
0.050105124711990356,
0.1518610119819641,
-0.11155825108289719,
-0.046714674681425095,
0.1081906408071518,
0.019937077537178993,
0.03223314508795738,
-0.027893399819731712,
-0.02081061340868473,
-0.009831633418798447,
-0.15140125155448914,
-0.06066659092903137,
0.32738298177719116,
0.0219104066491127,
-0.014895390719175339,
0.014952598139643669,
-0.07323130965232849,
-0.15754662454128265,
0.030887654051184654,
-0.06898327171802521,
-0.02699069119989872,
0.036231521517038345,
0.1487724483013153,
-0.087316133081913,
-0.1293564736843109,
-0.13637633621692657,
-0.0406554751098156,
0.21305841207504272,
0.028782803565263748,
0.03711417317390442,
-0.12386038899421692,
0.10282314568758011,
0.02528967335820198,
0.021479761227965355,
0.004575804341584444,
-0.08393687754869461,
0.028942639008164406,
-0.08229923248291016,
0.02840992622077465,
0.03957226127386093,
0.05314154177904129,
0.12975211441516876,
-0.24745428562164307,
0.0498918853700161,
-0.01221481803804636,
0.10228876024484634,
-0.08759955316781998,
-0.012934451922774315,
0.059870440512895584,
0.10758087038993835,
0.011062410660088062,
0.06503567099571228,
-0.04047826677560806,
0.04769480600953102,
-0.015642637386918068,
-0.017797106876969337,
0.04686347395181656,
0.14800938963890076,
-0.12264758348464966,
-0.046683818101882935,
-0.06526487320661545,
0.021177683025598526,
0.08030859380960464,
-0.10964316874742508,
-0.049426399171352386,
0.037315916270017624,
0.03680378943681717,
-0.052375100553035736,
0.04209700599312782,
-0.01810240000486374,
-0.029695948585867882,
0.10477819293737411,
-0.08702286332845688,
0.010991969145834446,
-0.06858786195516586,
0.032282840460538864,
-0.0012414238881319761,
-0.05746341496706009,
0.0034915502183139324,
-0.14875246584415436,
-0.13707521557807922,
-0.06146319955587387,
0.0289689302444458,
0.030304862186312675,
-0.001960750436410308,
-0.04644375666975975,
-0.0592380128800869,
-0.08154382556676865,
-0.06459696590900421,
-0.07681991159915924,
-0.05353395268321037,
0.09663441777229309,
-0.10969588905572891,
0.004872849211096764,
-0.09001162648200989,
-0.0032616762910038233,
-0.05085869878530502,
0.06292290985584259,
-0.0727708712220192,
-0.02992992289364338,
-0.0995783805847168,
0.11844996362924576,
-0.057386960834264755,
-0.08405909687280655,
-0.23416833579540253,
0.02355174534022808,
-0.013477092608809471,
0.21090209484100342,
-0.04458089917898178,
0.002140712458640337,
0.2927594482898712,
-0.1412450671195984,
-0.1215098425745964,
0.04224296659231186,
-0.0003904782060999423,
0.12281900644302368,
0.04208912327885628,
0.23480187356472015,
0.043084658682346344,
-0.23528136312961578,
0.1642545759677887,
0.11152003705501556,
-0.04149959236383438,
-0.10205233842134476,
0.041343022137880325,
-0.1310075968503952,
-0.1848418265581131,
-0.02203703485429287,
-0.07880376279354095,
0.07269614934921265,
-0.08756417781114578,
-0.03593392297625542,
-0.0008668560767546296,
-0.07305436581373215,
0.09011940658092499,
0.02526750974357128,
0.09444047510623932,
-0.10078554600477219,
0.04933442547917366,
0.09543810784816742,
0.06644144654273987,
0.06382283568382263,
-0.029553912580013275,
-0.02870125137269497,
-0.0034998366609215736,
-0.009989667683839798,
-0.049025822430849075,
-0.08604838699102402,
-0.1737370789051056,
-0.01942473277449608,
0.0748635083436966,
-0.09358280897140503,
0.07169465720653534,
0.1008106917142868,
0.07343634217977524,
-0.02662547305226326,
-0.08873605728149414,
-0.09369941800832748,
0.0349210686981678,
-0.04676230996847153,
-0.01619827002286911,
0.011747940443456173,
-0.053646255284547806,
0.12881991267204285,
-0.2166789174079895,
0.08161535114049911,
0.043580226600170135,
0.06568413227796555,
0.10245595127344131,
-0.04308951646089554,
0.012595201842486858,
0.07087333500385284,
0.04045867174863815,
-0.03426321595907211,
0.08352223038673401,
0.04118926450610161,
0.054313596338033676,
-0.035729147493839264,
-0.07493831217288971,
0.33381450176239014,
0.10556667298078537,
-0.01363906729966402,
-0.07217156141996384,
-0.09939087927341461,
-0.10779250413179398,
0.033311814069747925,
-0.07231571525335312,
0.011821071617305279,
0.028648022562265396,
0.0028891803231090307,
0.14613892138004303,
-0.06384652853012085,
-0.04092941805720329,
0.007764911279082298,
-0.08052564412355423,
0.0008381258812732995,
0.014457386918365955,
0.01594236120581627,
-0.013772484846413136,
0.1647508442401886,
0.15696734189987183,
0.014257015660405159,
0.17608021199703217,
-0.04961535707116127,
-0.05896269530057907,
0.03959165886044502,
0.2302849441766739,
-0.014607272110879421,
0.15490488708019257,
-0.055795613676309586,
0.029887542128562927,
-0.007573585491627455,
0.08190004527568817,
0.09629849344491959,
-0.21493035554885864,
-0.1131359189748764,
-0.03634883463382721,
-0.05067921057343483,
-0.08708503097295761,
0.042312536388635635,
0.012895507737994194,
0.09529579430818558,
0.0182382520288229,
-0.049761757254600525,
0.08861486613750458,
-0.0006361150881275535,
-0.11874070763587952,
0.14959020912647247,
-0.15184719860553741,
-0.21782442927360535,
-0.19954505562782288,
0.1290128231048584,
0.016229894012212753,
0.03709223121404648,
0.026668235659599304,
-0.15705853700637817,
0.013110472820699215,
-0.015511699952185154,
-0.02786264196038246,
-0.05091976001858711,
-0.056792665272951126,
0.05838988721370697,
0.10989327728748322,
0.044141873717308044,
-0.06923068314790726,
-0.032670725136995316,
-0.0346972793340683,
-0.10877672582864761,
0.054669078439474106,
-0.10706876963376999,
0.034730464220047,
0.10364020615816116,
0.013992248103022575,
0.08408955484628677,
-0.04886309430003166,
0.1734282225370407,
-0.10929463058710098,
-0.03453477472066879,
0.14357808232307434,
-0.06774876266717911,
0.019849296659231186,
0.07113417983055115,
0.034638721495866776,
-0.17141874134540558,
0.03973573073744774,
0.002701009390875697,
-0.09143519401550293,
-0.25087589025497437,
-0.08472536504268646,
-0.06476637721061707,
0.02185320109128952,
0.05633729323744774,
0.10783201456069946,
0.09162920713424683,
0.056402046233415604,
0.008337428793311119,
-0.1321275532245636,
0.059426408261060715,
0.013513035140931606,
-0.03561565279960632,
-0.056183166801929474,
0.02739715948700905,
-0.09186311066150665,
-0.04450976848602295,
0.06564988195896149,
0.07800836116075516,
0.2384876161813736,
0.13883930444717407,
-0.20824487507343292,
0.09098093211650848,
0.19917502999305725,
0.08418740332126617,
0.17487360537052155,
0.050074461847543716,
-0.043410297483205795,
0.014447673223912716,
-0.0122060002759099,
-0.11220796406269073,
0.06926801055669785,
-0.135521799325943,
0.054912012070417404,
-0.07433676719665527,
-0.15065467357635498,
0.08282385021448135,
0.3167039453983307,
0.00528634013608098,
-0.2313772588968277,
-0.05795358121395111,
0.03146638721227646,
0.043990060687065125,
-0.11362040787935257,
0.16652292013168335,
0.1819610297679901,
-0.1301470398902893,
0.018011964857578278,
-0.04059407860040665,
0.0676456168293953,
0.013950444757938385,
0.022686131298542023,
-0.017322653904557228,
-0.05053432285785675,
-0.01857675239443779,
0.04271508753299713,
-0.17481468617916107,
0.31005844473838806,
-0.022758854553103447,
0.08659613877534866,
0.03899727389216423,
-0.04240472987294197,
0.041567232459783554,
0.20657040178775787,
0.2175041288137436,
0.058718036860227585,
-0.011042571626603603,
-0.20268131792545319,
-0.03590092435479164,
0.013515752740204334,
0.1001753956079483,
-0.03411423787474632,
-0.04571685567498207,
-0.04346652328968048,
0.06428553909063339,
-0.019126955419778824,
-0.14174270629882812,
-0.12322108447551727,
-0.02038327045738697,
-0.08564576506614685,
0.020307479426264763,
0.11966709792613983,
-0.14677712321281433,
-0.0017738471506163478,
-0.04859147220849991,
0.03481116145849228,
-0.1394895762205124,
-0.02589433081448078,
-0.09254027903079987,
-0.18995021283626556,
-0.0043361796997487545,
0.017015468329191208,
0.04640854522585869,
-0.008712458424270153,
-0.003246455453336239,
-0.06386867165565491,
-0.043123453855514526,
0.10104965418577194,
-0.1383085548877716,
-0.021074626594781876,
-0.02880520559847355,
0.21791154146194458,
0.007281738333404064,
-0.02903485670685768,
-0.010748236440122128,
-0.0015381420962512493,
0.06665600836277008,
-0.09800155460834503,
0.02462327480316162,
0.06839415431022644,
0.004031648393720388,
0.04628736898303032,
-0.15946553647518158,
0.05590582638978958,
-0.03938259556889534,
-0.015455832704901695,
0.17522938549518585,
0.2399473339319229,
-0.06958426535129547,
0.05149579048156738,
0.06540724635124207,
-0.04382859915494919,
-0.25070127844810486,
0.06464546918869019,
-0.013721720315515995,
-0.018876787275075912,
0.009190880693495274,
-0.17024138569831848,
0.013110560365021229,
0.23010271787643433,
-0.06771379709243774,
0.17641562223434448,
-0.2858361005783081,
-0.016158485785126686,
0.13363422453403473,
0.09186282753944397,
0.2495933324098587,
-0.15584763884544373,
-0.057036299258470535,
0.06046558544039726,
-0.003053384367376566,
0.013369114138185978,
-0.20817133784294128,
0.010900712572038174,
0.010041081346571445,
-0.08714208006858826,
0.004193995613604784,
-0.025930272415280342,
0.2225557565689087,
-0.045666925609111786,
0.11154845356941223,
-0.05894266441464424,
0.026169393211603165,
-0.052125439047813416,
-0.0378970131278038,
0.061482034623622894,
0.11295752227306366,
0.03806766867637634,
-0.15290454030036926,
0.00040992084541358054,
-0.0646514892578125,
0.07761375606060028,
-0.03258838504552841,
-0.0549466498196125,
0.009096681140363216,
0.014751063659787178,
-0.05070741847157478,
0.03925710171461105,
0.13638603687286377,
-0.04843146726489067,
0.3315286934375763,
0.1182451993227005,
0.0636453777551651,
-0.10111770033836365,
-0.07467970252037048,
0.01022570300847292,
-0.07149922102689743,
0.11787120252847672,
-0.1243802011013031,
-0.008817701600492,
0.0657954141497612,
0.010608786717057228,
0.11464615911245346,
0.06962022930383682,
-0.01469535380601883,
-0.016169700771570206,
0.11055222153663635,
-0.2150479406118393,
-0.17720745503902435,
-0.04795600473880768,
0.05455835536122322,
0.05748480185866356,
0.10636111348867416,
0.10540628433227539,
-0.04839744046330452,
0.01577913947403431,
-0.058463290333747864,
0.006163392215967178,
-0.12024889141321182,
0.05779222026467323,
0.10494714975357056,
0.06800201535224915,
-0.0328449085354805,
0.08870799839496613,
0.02947908826172352,
-0.09688828140497208,
0.004283994901925325,
0.0970001071691513,
-0.10748915374279022,
-0.061016518622636795,
-0.038313839584589005,
0.08663873374462128,
0.12017454206943512,
-0.11058824509382248,
-0.02906147763133049,
-0.12327799946069717,
0.014911332167685032,
0.11742620915174484,
0.048426639288663864,
0.04520079866051674,
0.06155077740550041,
0.0364457331597805,
-0.03770521283149719,
-0.03296937793493271,
-0.0759870782494545,
0.030482077971100807,
-0.17871545255184174,
0.05469413101673126,
0.021059710532426834,
0.050592586398124695,
-0.06189772114157677,
-0.04030199721455574,
-0.14239755272865295,
0.08722434192895889,
0.0236711073666811,
-0.046740446239709854,
-0.055460553616285324,
-0.0020890417508780956,
0.000012282402167329565,
-0.038131244480609894,
-0.03783369064331055,
0.02104518562555313,
-0.11226243525743484,
0.08085237443447113,
0.007766183465719223,
0.045474305748939514,
-0.028526267036795616,
0.013892948627471924,
0.07019530981779099,
-0.07374338805675507,
0.0878622755408287,
0.058308299630880356,
-0.04674550145864487,
0.11017481237649918,
-0.08291705697774887,
0.016409756615757942,
0.14389705657958984,
-0.0646931454539299,
0.04200272262096405,
0.050961971282958984,
0.0008014820050448179,
-0.006602439563721418,
0.01695593073964119,
0.0804436132311821,
0.051423974335193634,
-0.10033298283815384,
-0.024841468781232834,
0.00622575031593442,
-0.10022774338722229,
-0.03522656112909317,
-0.08546700328588486,
0.14547961950302124,
0.016630610451102257,
0.11973689496517181,
-0.10593896359205246,
0.03359372541308403,
-0.10388454049825668,
-0.012829430401325226,
-0.016863172873854637,
-0.025347258895635605,
-0.02452128566801548,
0.008236419409513474,
0.023511098697781563,
-0.0027427596505731344,
0.16054002940654755,
-0.05436990037560463,
-0.016419939696788788,
0.049103058874607086,
-0.1436406672000885,
-0.05446966364979744,
0.0351385734975338,
0.18491658568382263,
0.10237450897693634,
-0.06315352767705917,
-0.07874928414821625,
0.016599904745817184,
-0.02094179578125477,
-0.1600126177072525,
0.1449240893125534,
0.16955387592315674,
-0.15771836042404175,
0.055356454104185104,
0.027398640289902687,
-0.042895715683698654,
0.0025541959330439568,
-0.05650975927710533,
0.054110053926706314,
-0.0474151112139225,
0.005609964951872826,
0.06282229721546173,
0.27388739585876465,
-0.12437301874160767,
0.02190467342734337,
-0.013217803090810776,
-0.03998721018433571,
-0.1706298142671585,
0.08399541676044464,
-0.028687475249171257,
-0.15854699909687042,
0.021642938256263733,
-0.06976169347763062,
-0.08597008138895035,
0.18508562445640564,
0.015926973894238472,
0.02956329844892025,
0.29571759700775146,
0.00032706125057302415,
-0.018923766911029816,
0.045247357338666916,
0.010524434968829155,
0.04432244226336479,
-0.18684953451156616,
-0.14228057861328125,
0.02507493644952774,
-0.12088751047849655,
-0.009391247294843197,
-0.034141018986701965,
-0.06717637181282043,
-0.002290202770382166,
-0.0772051215171814,
-0.006884424947202206,
-0.017468225210905075,
0.10394603759050369,
-0.11690228432416916,
0.0947209820151329,
0.05821515619754791,
-0.07654059678316116,
0.0018734013428911567,
0.16737854480743408,
-0.04256851226091385,
-0.07500255852937698,
-0.1368902623653412,
0.2745916545391083,
0.0352877639234066,
0.12906883656978607,
-0.03710036724805832,
0.00024492270313203335,
0.018200047314167023,
0.18095456063747406,
0.17571809887886047,
-0.03301902860403061,
0.045950207859277725,
-0.0038470111321657896,
0.027860265225172043,
-0.004076339770108461,
0.20277923345565796,
0.1283283233642578,
0.09978311508893967,
-0.03479553759098053,
0.005490999203175306,
-0.026528961956501007,
0.06848831474781036,
-0.0975690558552742,
0.017763204872608185,
0.027765505015850067,
0.01814870722591877,
-0.12497733533382416,
0.16609515249729156,
-0.05738289654254913,
0.0677499994635582,
0.09404187649488449,
-0.11205779761075974,
-0.10587383061647415,
-0.022465374320745468,
0.01777215115725994,
-0.07975313812494278,
0.0695631355047226,
-0.059780336916446686,
-0.09321082383394241,
0.053116366267204285,
0.01062578335404396,
-0.05980269983410835,
-0.21117016673088074,
0.04179225116968155,
-0.006525810807943344,
0.2227240800857544,
0.004721134901046753,
-0.00851003173738718,
0.02078585885465145,
-0.059220168739557266,
-0.05172068253159523,
0.13469600677490234,
0.04916968569159508,
0.04595471918582916,
-0.13025569915771484,
-0.018766408786177635,
-0.051412031054496765,
0.0348595529794693,
0.033246539533138275,
-0.20658951997756958,
-0.009480167180299759,
-0.0031631444580852985,
-0.11438208073377609,
-0.07636937499046326,
0.07887046039104462,
-0.043010544031858444,
0.09708095341920853,
0.10506469756364822,
-0.003957798238843679,
0.039706021547317505,
-0.05389934033155441,
0.11468890309333801,
0.07101957499980927,
-0.04775277152657509,
-0.025733033195137978,
-0.1291586011648178,
-0.017691364511847496,
-0.03080946020781994,
-0.08023548126220703,
-0.15576772391796112,
-0.022709952667355537,
-0.07977133989334106,
0.01766781136393547,
-0.09350316971540451,
0.022238900884985924,
0.0018009408377110958,
0.05920787528157234,
0.007994246669113636,
-0.17437966167926788,
0.003941001370549202,
0.12216940522193909,
0.015576873905956745,
-0.04441463574767113
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# PXAudio Whisper For user_476da26872df492f830a65925d422651
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the ja 0.1 dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 50
### Training results
### Framework versions
- Transformers 4.33.3
- Pytorch 2.2.0+cu118
- Datasets 2.16.1
- Tokenizers 0.13.3
| {"language": ["ja"], "license": "apache-2.0", "tags": ["hf-asr-leaderboard", "generated_from_trainer"], "datasets": ["pxaudio/user_476da26872df492f830a65925d422651_model"], "base_model": "openai/whisper-small", "model-index": [{"name": "PXAudio Whisper For user_476da26872df492f830a65925d422651", "results": []}]} | automatic-speech-recognition | hoangvanvietanh/user_476da26872df492f830a65925d422651_model | [
"transformers",
"pytorch",
"whisper",
"automatic-speech-recognition",
"hf-asr-leaderboard",
"generated_from_trainer",
"ja",
"dataset:pxaudio/user_476da26872df492f830a65925d422651_model",
"base_model:openai/whisper-small",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-06T08:27:35+00:00 | [] | [
"ja"
] | TAGS
#transformers #pytorch #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #ja #dataset-pxaudio/user_476da26872df492f830a65925d422651_model #base_model-openai/whisper-small #license-apache-2.0 #endpoints_compatible #region-us
|
# PXAudio Whisper For user_476da26872df492f830a65925d422651
This model is a fine-tuned version of openai/whisper-small on the ja 0.1 dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 50
### Training results
### Framework versions
- Transformers 4.33.3
- Pytorch 2.2.0+cu118
- Datasets 2.16.1
- Tokenizers 0.13.3
| [
"# PXAudio Whisper For user_476da26872df492f830a65925d422651\n\nThis model is a fine-tuned version of openai/whisper-small on the ja 0.1 dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- training_steps: 50",
"### Training results",
"### Framework versions\n\n- Transformers 4.33.3\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #ja #dataset-pxaudio/user_476da26872df492f830a65925d422651_model #base_model-openai/whisper-small #license-apache-2.0 #endpoints_compatible #region-us \n",
"# PXAudio Whisper For user_476da26872df492f830a65925d422651\n\nThis model is a fine-tuned version of openai/whisper-small on the ja 0.1 dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- training_steps: 50",
"### Training results",
"### Framework versions\n\n- Transformers 4.33.3\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.13.3"
] | [
106,
54,
6,
12,
8,
3,
104,
4,
33
] | [
"passage: TAGS\n#transformers #pytorch #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #ja #dataset-pxaudio/user_476da26872df492f830a65925d422651_model #base_model-openai/whisper-small #license-apache-2.0 #endpoints_compatible #region-us \n# PXAudio Whisper For user_476da26872df492f830a65925d422651\n\nThis model is a fine-tuned version of openai/whisper-small on the ja 0.1 dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- training_steps: 50### Training results### Framework versions\n\n- Transformers 4.33.3\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.13.3"
] | [
-0.12377674877643585,
0.18442083895206451,
-0.0038589853793382645,
0.09028810262680054,
0.10626311600208282,
0.0285260621458292,
0.10860702395439148,
0.17101843655109406,
-0.038902658969163895,
0.10248441249132156,
0.06980402767658234,
0.040130238980054855,
0.07975504547357559,
0.1689399629831314,
-0.03384440019726753,
-0.2609032392501831,
-0.008135715499520302,
-0.03488800302147865,
-0.06800241023302078,
0.07802745699882507,
0.1105327233672142,
-0.0861852839589119,
0.056361567229032516,
0.0028491313569247723,
-0.0966232419013977,
0.012678717263042927,
-0.03822575882077217,
-0.06084546819329262,
0.06523676961660385,
0.014853041619062424,
0.04653296247124672,
0.008906194940209389,
0.10953346639871597,
-0.21512232720851898,
-0.009006163105368614,
0.05365002155303955,
0.04908793419599533,
0.0858994647860527,
0.09147833287715912,
0.012945301830768585,
0.05857876315712929,
-0.14341405034065247,
0.08510120958089828,
0.05217278376221657,
-0.0708458423614502,
-0.1676332801580429,
-0.08238192647695541,
0.10204782336950302,
0.09095141291618347,
0.09396929293870926,
-0.0037985509261488914,
0.07911353558301926,
-0.08783785998821259,
0.07071608304977417,
0.17711414396762848,
-0.2154017835855484,
-0.06206613779067993,
0.02285829372704029,
0.013227047398686409,
0.04375266656279564,
-0.1266936957836151,
-0.02674371749162674,
0.029270291328430176,
0.004349556751549244,
0.0692623108625412,
-0.004817286506295204,
-0.01529929880052805,
-0.009668403305113316,
-0.09718702733516693,
-0.03521076589822769,
0.15456895530223846,
0.06733206659555435,
-0.044117312878370285,
-0.1821880042552948,
-0.042517680674791336,
-0.0987229272723198,
-0.030273061245679855,
-0.008647208102047443,
0.03793404623866081,
-0.05082513019442558,
-0.01794624887406826,
-0.02201923355460167,
-0.050439272075891495,
-0.07005523145198822,
0.042046919465065,
0.1126440167427063,
0.03277675434947014,
0.002390787238255143,
-0.015330939553678036,
0.08233968168497086,
0.017236260697245598,
-0.14698077738285065,
-0.03960380703210831,
0.010552916675806046,
-0.11285664886236191,
-0.03237103298306465,
-0.003986569121479988,
-0.012542436830699444,
0.00827151257544756,
0.140024334192276,
-0.026555076241493225,
0.08795365691184998,
0.03740083426237106,
-0.004366623237729073,
-0.010131713934242725,
0.15945813059806824,
-0.05181040242314339,
-0.06436479836702347,
-0.025228850543498993,
0.12734785676002502,
0.008374354802072048,
-0.014371968805789948,
-0.06659476459026337,
-0.046926286071538925,
0.050555095076560974,
0.07008926570415497,
-0.0037936612498015165,
0.016107315197587013,
-0.05390466749668121,
-0.049212291836738586,
0.03224623203277588,
-0.1518283635377884,
0.044861629605293274,
0.021968603134155273,
-0.09611144661903381,
-0.011670888401567936,
0.0001888744009193033,
-0.005535339470952749,
-0.05640720948576927,
0.0861923098564148,
-0.06163879111409187,
-0.03326046094298363,
-0.06314641237258911,
-0.034463103860616684,
0.006888668052852154,
-0.09039356559515,
0.004718576092272997,
-0.05380202829837799,
-0.18937931954860687,
-0.06833282113075256,
0.031915292143821716,
-0.08287778496742249,
-0.08344574272632599,
-0.05447430536150932,
-0.07589466869831085,
0.03638510778546333,
-0.00102095992770046,
0.14015832543373108,
-0.03736557066440582,
0.06449075788259506,
-0.0001285848265979439,
0.04374907538294792,
0.07774341851472855,
0.04728628695011139,
-0.04508470743894577,
0.04213287681341171,
-0.10045621544122696,
0.11936471611261368,
-0.10488645732402802,
0.03094841167330742,
-0.13096126914024353,
-0.09615158289670944,
-0.0004887853865511715,
-0.013807086274027824,
0.054000046104192734,
0.13948240876197815,
-0.1720670610666275,
-0.03410335257649422,
0.1531168669462204,
-0.04741470143198967,
-0.0604727566242218,
0.07805126160383224,
-0.026482829824090004,
0.02090149000287056,
0.03157621994614601,
0.18046978116035461,
0.16038759052753448,
-0.12284837663173676,
-0.012847712263464928,
0.026633068919181824,
0.0582934208214283,
0.04039856791496277,
0.07237636297941208,
-0.03655286505818367,
0.06428256630897522,
0.01256993692368269,
-0.07656442373991013,
0.0025103187654167414,
-0.045704323798418045,
-0.08356252312660217,
-0.02790883183479309,
-0.09609291702508926,
0.02058013714849949,
0.033587515354156494,
0.031261149793863297,
-0.05484432354569435,
-0.10143735259771347,
0.036616332828998566,
0.16799819469451904,
-0.04549405723810196,
-0.011495509184896946,
-0.11072352528572083,
-0.0025254564825445414,
0.017260504886507988,
-0.02196190319955349,
-0.14893142879009247,
-0.03408454731106758,
0.050816282629966736,
-0.09831420332193375,
-0.003661535680294037,
0.015775199979543686,
0.08576846867799759,
0.03643222525715828,
-0.03796396404504776,
-0.05979090556502342,
-0.061790984123945236,
-0.02114800550043583,
-0.07388217747211456,
-0.17936061322689056,
-0.08343755453824997,
-0.014448997564613819,
0.19160272181034088,
-0.20826363563537598,
0.012722798623144627,
0.02265908755362034,
0.13309229910373688,
0.03904663398861885,
-0.062395427376031876,
0.05046408995985985,
0.019284294918179512,
0.028553897514939308,
-0.1102210208773613,
0.018856942653656006,
-0.00203415984287858,
-0.08323130756616592,
-0.03414417803287506,
-0.1151924580335617,
0.050523027777671814,
0.05087462067604065,
0.13185562193393707,
-0.08698773384094238,
-0.00629032775759697,
-0.05678457021713257,
-0.046313561499118805,
-0.06302442401647568,
-0.008246601559221745,
0.1998874992132187,
0.028436435386538506,
0.11204899847507477,
-0.06618965417146683,
-0.08246134966611862,
-0.0018031388754025102,
0.02174355275928974,
-0.005796991754323244,
0.12476690858602524,
0.0303014125674963,
-0.08520174771547318,
0.07352880388498306,
0.036585692316293716,
-0.0304380152374506,
0.17194634675979614,
-0.0636289194226265,
-0.09664501249790192,
-0.028759216889739037,
0.04453381150960922,
0.016779690980911255,
0.10745663195848465,
-0.07429049909114838,
0.004482825752347708,
0.04842853918671608,
0.008999002166092396,
0.024131035432219505,
-0.12205912917852402,
-0.007362340111285448,
0.0444791354238987,
-0.04778173565864563,
0.0002949631598312408,
-0.015045891515910625,
0.027283400297164917,
0.07868557423353195,
0.03796926513314247,
-0.014898337423801422,
0.0016933426959440112,
-0.03744954615831375,
-0.08111081272363663,
0.16187086701393127,
-0.09356755763292313,
-0.14259867370128632,
-0.12643903493881226,
0.05522023141384125,
-0.01752672716975212,
-0.02098778821527958,
-0.017267199233174324,
-0.09207817167043686,
-0.05301272124052048,
-0.07936301082372665,
-0.026022804901003838,
-0.03012883849442005,
-0.01629459485411644,
0.09178008884191513,
0.03950067609548569,
0.10146747529506683,
-0.1245332881808281,
0.020330827683210373,
0.0066886176355183125,
-0.07625385373830795,
-0.0077165598049759865,
0.04271519184112549,
0.07648701220750809,
0.11604022234678268,
0.019492581486701965,
0.03102053701877594,
-0.023524541407823563,
0.19701701402664185,
-0.10633399337530136,
0.021265478804707527,
0.1346207708120346,
0.03151202201843262,
0.05357677489519119,
0.12658973038196564,
0.025247013196349144,
-0.08374939858913422,
0.01168984454125166,
0.0442771352827549,
0.001955188112333417,
-0.22820229828357697,
-0.028182940557599068,
-0.04044660925865173,
0.003427867777645588,
0.1428479254245758,
0.053559932857751846,
-0.009302684105932713,
0.05520033836364746,
-0.017352886497974396,
0.03678039833903313,
-0.015083270147442818,
0.06540104746818542,
0.0366092287003994,
0.03656211495399475,
0.08125866204500198,
-0.01932676136493683,
-0.0003514296840876341,
0.07079767435789108,
0.0029798345640301704,
0.1775440275669098,
-0.058615606278181076,
0.19289669394493103,
0.00017437916540075094,
0.14785324037075043,
-0.012857986614108086,
0.04121093079447746,
0.021886533126235008,
-0.004174773581326008,
0.022697294130921364,
-0.07487356662750244,
-0.06826060265302658,
0.05137965828180313,
0.03502153977751732,
0.048626624047756195,
-0.07787590473890305,
0.06934988498687744,
0.007405930198729038,
0.2887190282344818,
0.05142388865351677,
-0.2588382661342621,
-0.07351415604352951,
0.019421955570578575,
-0.058775316923856735,
-0.0823267325758934,
0.012513462454080582,
0.14438076317310333,
-0.15634490549564362,
0.05972977727651596,
-0.03647102788090706,
0.08794676512479782,
-0.08733860403299332,
-0.012000542134046555,
0.05665012076497078,
0.10095804929733276,
0.00785457156598568,
0.07866260409355164,
-0.18811163306236267,
0.17947107553482056,
0.003785739652812481,
0.10336070507764816,
-0.06019636243581772,
0.053530361503362656,
0.01947658136487007,
-0.004672950599342585,
0.13025151193141937,
0.0024960681330412626,
-0.04302003234624863,
-0.16150537133216858,
-0.10562548041343689,
0.012086148373782635,
0.11002537608146667,
-0.08303368091583252,
0.07776587456464767,
-0.04764774441719055,
-0.023850811645388603,
0.005193660967051983,
-0.08730793744325638,
-0.10696402937173843,
-0.13260720670223236,
0.03759036958217621,
0.027846580371260643,
0.0006136744632385671,
-0.09730605036020279,
-0.07126907259225845,
0.005119217559695244,
0.1307031214237213,
-0.0856117457151413,
-0.07049982994794846,
-0.15938860177993774,
0.045279573649168015,
0.15069355070590973,
-0.07296914607286453,
0.02956298179924488,
0.041327930986881256,
0.1544414758682251,
0.028269577771425247,
-0.047458402812480927,
0.035547494888305664,
-0.07847285270690918,
-0.19198016822338104,
-0.041651420295238495,
0.15810593962669373,
0.026917386800050735,
0.055744461715221405,
-0.0116459671407938,
0.013642504811286926,
0.019144628196954727,
-0.07988696545362473,
0.04686877876520157,
0.0925593227148056,
0.02125018648803234,
0.059542011469602585,
-0.03243410959839821,
0.019033584743738174,
-0.04606039822101593,
-0.005388992372900248,
0.1100238487124443,
0.20917868614196777,
-0.08401942253112793,
0.0911155492067337,
0.042994603514671326,
-0.059743136167526245,
-0.1572646200656891,
0.03104345127940178,
0.14312900602817535,
0.04478674754500389,
0.01605191081762314,
-0.22749446332454681,
0.08824733644723892,
0.08865831047296524,
-0.028135299682617188,
0.041722893714904785,
-0.33610793948173523,
-0.12390606105327606,
0.07428038865327835,
0.06618417799472809,
-0.051736123859882355,
-0.12254228442907333,
-0.06837687641382217,
-0.03767671808600426,
-0.13586090505123138,
0.018826933577656746,
-0.003528625937178731,
0.10357069224119186,
-0.029861170798540115,
0.08081664144992828,
0.05533437058329582,
-0.032287079840898514,
0.13159684836864471,
0.02882852777838707,
0.059883762151002884,
-0.06212982162833214,
0.0069208815693855286,
0.015928873792290688,
-0.07642567902803421,
0.06838452816009521,
-0.0608147531747818,
0.0725955218076706,
-0.19746696949005127,
-0.02117115631699562,
-0.07407566905021667,
0.03139360621571541,
-0.04818947613239288,
-0.05558190494775772,
-0.019771931692957878,
0.05416957661509514,
0.07601799815893173,
-0.008920480497181416,
-0.0035245923791080713,
-0.0006544449715875089,
0.026493633165955544,
0.12037701159715652,
0.1234142929315567,
0.020954033359885216,
-0.15394407510757446,
-0.017968030646443367,
0.00901057105511427,
0.05784573405981064,
-0.1197146400809288,
0.03058629482984543,
0.11117357760667801,
0.07243244349956512,
0.11936691403388977,
0.018967390060424805,
-0.059868745505809784,
-0.02060018852353096,
0.03370041772723198,
-0.07982691377401352,
-0.1598404496908188,
-0.024148382246494293,
0.039278604090213776,
-0.151199609041214,
-0.03689264878630638,
0.09980250895023346,
-0.03572801128029823,
-0.01459898054599762,
-0.01168773788958788,
0.034309010952711105,
0.0021299864165484905,
0.17854951322078705,
0.049127135425806046,
0.09262139350175858,
-0.07743281871080399,
0.11393903940916061,
0.07296653836965561,
-0.04985538870096207,
0.07236770540475845,
0.10422341525554657,
-0.06009730324149132,
-0.004994349554181099,
0.04796259477734566,
0.09332685172557831,
0.03522631525993347,
-0.035852015018463135,
-0.05246417969465256,
-0.10577265918254852,
0.06785073131322861,
0.0356028787791729,
0.02480660378932953,
-0.004721315111964941,
-0.029956670477986336,
0.010424789041280746,
-0.13055455684661865,
0.08783204853534698,
0.043318428099155426,
0.04530671238899231,
-0.15036438405513763,
0.10691607743501663,
-0.01902594417333603,
0.018672635778784752,
0.004034295678138733,
0.014441333711147308,
-0.08485067635774612,
-0.00949693564325571,
-0.08573982119560242,
0.02738931216299534,
-0.026189975440502167,
0.00153851299546659,
-0.024558771401643753,
-0.03568374365568161,
-0.026624608784914017,
0.04895215481519699,
-0.08110594004392624,
-0.07978838682174683,
-0.0025817533023655415,
0.07302748411893845,
-0.0996326133608818,
-0.0022585191763937473,
0.03624367341399193,
-0.11728864163160324,
0.08863312005996704,
0.04774273931980133,
0.011001395992934704,
-0.0031507404055446386,
-0.04956161603331566,
-0.03562401607632637,
0.0029862264636904,
0.017209000885486603,
0.05613299086689949,
-0.14049315452575684,
-0.007056203670799732,
-0.04944575950503349,
0.004929449409246445,
0.0032910581212490797,
0.04191428795456886,
-0.13293498754501343,
0.019747866317629814,
-0.021042320877313614,
-0.03801148384809494,
-0.09231314063072205,
0.06727220863103867,
0.09952377527952194,
0.004435297101736069,
0.12276986241340637,
-0.058177024126052856,
0.06193360686302185,
-0.21220600605010986,
-0.023003334179520607,
-0.00044055155012756586,
-0.0139002101495862,
-0.04370143637061119,
-0.0055151511915028095,
0.09208294004201889,
-0.05395501106977463,
0.06793061643838882,
-0.03411433845758438,
0.007737298030406237,
0.02685164473950863,
-0.0789700299501419,
0.012920547276735306,
0.028290316462516785,
0.19342049956321716,
0.05825580283999443,
-0.02717866189777851,
0.11539905518293381,
-0.025849509984254837,
0.03036169894039631,
0.08786743134260178,
0.11864306032657623,
0.18368706107139587,
-0.0006652435404248536,
0.06445494294166565,
0.05424165725708008,
-0.13482002913951874,
-0.14673127233982086,
0.14046695828437805,
-0.07154561579227448,
0.11029519885778427,
-0.06969974935054779,
0.14860092103481293,
0.09361350536346436,
-0.18338094651699066,
0.07256671041250229,
-0.050029776990413666,
-0.10145256668329239,
-0.10587620735168457,
-0.09200853109359741,
-0.08208590745925903,
-0.09785552322864532,
0.029448525980114937,
-0.10864990949630737,
0.047434911131858826,
0.05499213933944702,
0.011631767265498638,
0.019652586430311203,
0.14254236221313477,
-0.06150169298052788,
0.006819551344960928,
0.10550215095281601,
0.01504588034003973,
-0.014778624288737774,
-0.061520323157310486,
-0.05729157105088234,
0.04573797062039375,
0.00990169309079647,
0.08656028658151627,
-0.05420123040676117,
-0.03394589200615883,
0.01457962766289711,
0.010955924168229103,
-0.0667477399110794,
0.030216164886951447,
-0.01436691451817751,
0.027768217027187347,
0.03604038432240486,
0.05640770494937897,
0.003431259421631694,
-0.046954505145549774,
0.2848023772239685,
-0.07962573319673538,
-0.08248112350702286,
-0.13866011798381805,
0.13150626420974731,
0.026895273476839066,
-0.015140109695494175,
0.06193796545267105,
-0.09654010087251663,
-0.015462630428373814,
0.15608811378479004,
0.1407497376203537,
-0.06406848132610321,
-0.015394068323075771,
-0.00899368617683649,
-0.0068292804062366486,
-0.05238775163888931,
0.08950489014387131,
0.07858385890722275,
0.040513865649700165,
-0.03906441479921341,
0.03968579322099686,
0.03440873324871063,
-0.07156594842672348,
-0.06248553842306137,
0.10978780686855316,
0.007431927137076855,
0.008617161773145199,
-0.044625941663980484,
0.06701205670833588,
-0.010488391853868961,
-0.1725032776594162,
0.04550769552588463,
-0.16407187283039093,
-0.1992107778787613,
-0.03173769637942314,
0.05008772760629654,
0.018346276134252548,
0.067184217274189,
0.014544686302542686,
-0.01763402484357357,
0.11623530834913254,
-0.009493283927440643,
-0.05350998789072037,
-0.0758686438202858,
0.06647473573684692,
-0.11866632848978043,
0.2200779914855957,
-0.007134847342967987,
0.03932235389947891,
0.1083412766456604,
0.007036656141281128,
-0.15549224615097046,
0.008266416378319263,
0.08295849710702896,
-0.07745937258005142,
0.03570304438471794,
0.21497590839862823,
-0.04612357169389725,
0.1121169850230217,
0.045567117631435394,
-0.13322213292121887,
-0.030794337391853333,
-0.01606232300400734,
0.008876267820596695,
-0.10331156849861145,
0.009006577543914318,
-0.0541381761431694,
0.14963287115097046,
0.2309814691543579,
-0.05906890705227852,
-0.029817378148436546,
-0.06077699735760689,
0.02987266704440117,
0.03562116622924805,
0.07503359019756317,
-0.026516661047935486,
-0.16985651850700378,
0.008505627512931824,
-0.02782633528113365,
0.0404462032020092,
-0.19201523065567017,
-0.08255157619714737,
0.02607574127614498,
-0.0583413690328598,
-0.033326540142297745,
0.11844854801893234,
0.047072410583496094,
0.041253477334976196,
-0.023162776604294777,
-0.10070068389177322,
-0.027561919763684273,
0.13952073454856873,
-0.14549273252487183,
-0.04190845042467117
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | keisar/publick1NVIDIAL4GPULeo | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-06T08:28:20+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SciBERT_25K_steps_bs64
This model is a fine-tuned version of [allenai/scibert_scivocab_uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0177
- Accuracy: 0.9941
- Precision: 0.7990
- Recall: 0.5288
- F1: 0.6364
- Hamming: 0.0059
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 25000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Hamming |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 0.0467 | 0.16 | 5000 | 0.0416 | 0.9902 | 0.0 | 0.0 | 0.0 | 0.0098 |
| 0.0236 | 0.32 | 10000 | 0.0223 | 0.9932 | 0.8192 | 0.3929 | 0.5311 | 0.0068 |
| 0.0198 | 0.47 | 15000 | 0.0190 | 0.9939 | 0.8015 | 0.4934 | 0.6108 | 0.0061 |
| 0.0185 | 0.63 | 20000 | 0.0180 | 0.9940 | 0.7974 | 0.5220 | 0.6310 | 0.0060 |
| 0.0181 | 0.79 | 25000 | 0.0177 | 0.9941 | 0.7990 | 0.5288 | 0.6364 | 0.0059 |
### Framework versions
- Transformers 4.35.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.7.1
- Tokenizers 0.14.1
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy", "precision", "recall", "f1"], "base_model": "allenai/scibert_scivocab_uncased", "model-index": [{"name": "SciBERT_25K_steps_bs64", "results": []}]} | text-classification | bdpc/SciBERT_25K_steps_bs64 | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:allenai/scibert_scivocab_uncased",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-06T08:32:03+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us
| SciBERT\_25K\_steps\_bs64
=========================
This model is a fine-tuned version of allenai/scibert\_scivocab\_uncased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0177
* Accuracy: 0.9941
* Precision: 0.7990
* Recall: 0.5288
* F1: 0.6364
* Hamming: 0.0059
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 25000
### Training results
### Framework versions
* Transformers 4.35.0.dev0
* Pytorch 2.0.1+cu118
* Datasets 2.7.1
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
61,
116,
4,
36
] | [
"passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000### Training results### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
-0.12256453931331635,
0.06859574466943741,
-0.0022490574046969414,
0.13081152737140656,
0.17255432903766632,
0.03548503667116165,
0.11826042830944061,
0.13099230825901031,
-0.06966569274663925,
0.03890903666615486,
0.12485005706548691,
0.144977867603302,
0.021586263552308083,
0.13492998480796814,
-0.05476110428571701,
-0.2883145809173584,
-0.0030204933136701584,
0.017089935019612312,
-0.06678253412246704,
0.13226324319839478,
0.08283255994319916,
-0.13384908437728882,
0.09503860771656036,
-0.020302239805459976,
-0.1611335277557373,
-0.0010213659843429923,
0.015837088227272034,
-0.04413985460996628,
0.14741744101047516,
0.01973525620996952,
0.12324615567922592,
0.022265395149588585,
0.11995978653430939,
-0.1749218851327896,
0.008455773815512657,
0.047751471400260925,
0.023260295391082764,
0.09193865209817886,
0.05296146124601364,
-0.0006148461252450943,
0.11154431104660034,
-0.09113859385251999,
0.05875547230243683,
0.020774967968463898,
-0.13279591500759125,
-0.17082759737968445,
-0.07799632847309113,
0.02885439619421959,
0.07418988645076752,
0.08998695760965347,
-0.006527826655656099,
0.10118331760168076,
-0.08838078379631042,
0.11019555479288101,
0.22057761251926422,
-0.2591956555843353,
-0.07662351429462433,
0.01093672402203083,
0.0036693087313324213,
0.09555799514055252,
-0.1072721853852272,
-0.011517575941979885,
0.03509452939033508,
0.03446045145392418,
0.12174569070339203,
-0.027128154411911964,
-0.12888427078723907,
0.022808870300650597,
-0.13819503784179688,
-0.02841782197356224,
0.09890762716531754,
0.02387155033648014,
-0.043325331062078476,
-0.027599472552537918,
-0.07610749453306198,
-0.17728374898433685,
-0.04193234071135521,
-0.011764860711991787,
0.03492835536599159,
-0.044703319668769836,
-0.07342393696308136,
-0.0255176592618227,
-0.08395376801490784,
-0.0968744233250618,
-0.04066105931997299,
0.17162592709064484,
0.03993697464466095,
0.012393137440085411,
-0.001596816349774599,
0.12435922026634216,
0.014813951216638088,
-0.15209978818893433,
0.009927665814757347,
0.01485131774097681,
-0.04943627491593361,
-0.04607813432812691,
-0.05734524130821228,
-0.015987347811460495,
-0.01900082267820835,
0.12946762144565582,
-0.029309451580047607,
0.05115506798028946,
0.03459009900689125,
0.01629892736673355,
-0.10418577492237091,
0.19656282663345337,
-0.05866871029138565,
-0.02264474146068096,
-0.00800324510782957,
0.08973747491836548,
0.007147806230932474,
-0.033029090613126755,
-0.10877956449985504,
-0.015745850279927254,
0.10305053740739822,
0.023373844102025032,
-0.08378916233778,
0.07007371634244919,
-0.04314819723367691,
-0.033242661505937576,
0.01964312233030796,
-0.09912638366222382,
0.035464946180582047,
0.01775357313454151,
-0.10126851499080658,
-0.01841958984732628,
0.024703437462449074,
0.004111442249268293,
-0.015314877033233643,
0.1283155232667923,
-0.09861647337675095,
0.034476783126592636,
-0.09383311867713928,
-0.12173494696617126,
-0.00561631890013814,
-0.09728076308965683,
0.0013282970758154988,
-0.08142926543951035,
-0.19219008088111877,
-0.029132945463061333,
0.0389828234910965,
-0.05216911435127258,
-0.03487739711999893,
-0.06878747045993805,
-0.06713918596506119,
0.02783927135169506,
-0.007858233526349068,
0.12839849293231964,
-0.05735670402646065,
0.11653177440166473,
0.03993707522749901,
0.07005761563777924,
-0.026736760511994362,
0.05856667086482048,
-0.10253626853227615,
0.011013761162757874,
-0.18385718762874603,
0.0675460621714592,
-0.05534125119447708,
0.04346661642193794,
-0.08069781213998795,
-0.10444338619709015,
0.024210505187511444,
0.0011464564595371485,
0.09312634915113449,
0.1407431960105896,
-0.1825769990682602,
-0.07279002666473389,
0.14212583005428314,
-0.068753182888031,
-0.0910632461309433,
0.11814436316490173,
-0.06597302854061127,
0.031408339738845825,
0.06318649649620056,
0.15239934623241425,
0.08369521051645279,
-0.06698165833950043,
0.015629073604941368,
-0.009996503591537476,
0.0834893211722374,
-0.01213387306779623,
0.07075818628072739,
0.02061934396624565,
-0.0012574979336932302,
0.025270503014326096,
-0.06098247691988945,
0.04153396561741829,
-0.10835004597902298,
-0.0941322073340416,
-0.03539029508829117,
-0.10403845459222794,
0.08269403129816055,
0.07324681431055069,
0.07785934209823608,
-0.08625990897417068,
-0.0879419669508934,
0.05588430166244507,
0.09884215891361237,
-0.06281279027462006,
0.0191056951880455,
-0.06192614883184433,
0.05420148745179176,
-0.023028796538710594,
-0.024973994120955467,
-0.18772950768470764,
-0.04566212370991707,
0.020164594054222107,
0.04873199015855789,
0.02279382385313511,
0.0014409752329811454,
0.08493226766586304,
0.08779004216194153,
-0.06492984294891357,
-0.030893953517079353,
-0.04037986695766449,
-0.008560266345739365,
-0.14008378982543945,
-0.20141084492206573,
-0.06439442187547684,
-0.019029079005122185,
0.13599766790866852,
-0.21441267430782318,
0.03438189998269081,
-0.014131131581962109,
0.08531447499990463,
0.019508324563503265,
-0.024535758420825005,
-0.032775893807411194,
0.07764223963022232,
-0.036688193678855896,
-0.05735108256340027,
0.06724438816308975,
-0.013345970772206783,
-0.0838085189461708,
-0.05072803050279617,
-0.11696246266365051,
0.13425147533416748,
0.10416813939809799,
-0.08975174278020859,
-0.09790020436048508,
0.004799272399395704,
-0.050574615597724915,
-0.031842418015003204,
-0.050319161266088486,
0.02574600838124752,
0.17270436882972717,
0.005644547753036022,
0.15299248695373535,
-0.05588493123650551,
-0.03578493371605873,
0.021297382190823555,
-0.007525456137955189,
0.04230661317706108,
0.13351191580295563,
0.1057264432311058,
-0.08503517508506775,
0.12989871203899384,
0.10885176807641983,
-0.08729502558708191,
0.1421014666557312,
-0.02460804395377636,
-0.08275175839662552,
-0.013688101433217525,
-0.02666560746729374,
0.001051764003932476,
0.10422473400831223,
-0.11482146382331848,
-0.018317507579922676,
0.011357538402080536,
0.014315959066152573,
-0.0036965636536478996,
-0.2061651200056076,
-0.03528577461838722,
0.041074588894844055,
-0.05428113043308258,
-0.033651940524578094,
-0.01438178215175867,
0.00835711881518364,
0.12099796533584595,
0.011896335519850254,
-0.09095581620931625,
0.013477628119289875,
-0.0025961154606193304,
-0.068870410323143,
0.21066002547740936,
-0.072234146296978,
-0.12254239618778229,
-0.12157904356718063,
-0.05403011292219162,
-0.058340031653642654,
0.028669031336903572,
0.03161579370498657,
-0.09426061064004898,
-0.009616715833544731,
-0.061849407851696014,
0.025041520595550537,
0.0050553916953504086,
0.04219931364059448,
-0.023693474009633064,
-0.020194532349705696,
0.058290764689445496,
-0.09001785516738892,
-0.003164287656545639,
-0.0592656247317791,
-0.08147405087947845,
0.03456597030162811,
0.030613062903285027,
0.11553838104009628,
0.15154269337654114,
-0.023628702387213707,
0.012759208679199219,
-0.02444758079946041,
0.22566962242126465,
-0.07143451273441315,
-0.00731214415282011,
0.1376505345106125,
-0.025471018627285957,
0.04772014543414116,
0.13169845938682556,
0.06357592344284058,
-0.08408086746931076,
0.010725535452365875,
0.043124012649059296,
-0.034448858350515366,
-0.2106553316116333,
-0.03699992597103119,
-0.05288170650601387,
0.006285845767706633,
0.10236874222755432,
0.014056635089218616,
0.02062990888953209,
0.07116186618804932,
0.016794173046946526,
0.04770670086145401,
-0.03809412568807602,
0.05932905897498131,
0.09631617367267609,
0.04400869458913803,
0.13809487223625183,
-0.025337180122733116,
-0.06682363152503967,
0.04239073768258095,
-0.02312398888170719,
0.20113560557365417,
-0.025888387113809586,
0.12194062024354935,
0.014655854552984238,
0.15776555240154266,
-0.009573941119015217,
0.08289414644241333,
0.0111967409029603,
-0.035880930721759796,
-0.011918353848159313,
-0.040099453181028366,
-0.06454755365848541,
0.02416330948472023,
-0.06219170242547989,
0.060528919100761414,
-0.15244939923286438,
0.014829927124083042,
0.047454651445150375,
0.27031829953193665,
0.06636649370193481,
-0.32915905117988586,
-0.10972490906715393,
0.0067866467870771885,
-0.03610401973128319,
-0.03386131674051285,
0.00924979243427515,
0.12802204489707947,
-0.10062500834465027,
0.02645345963537693,
-0.06910364329814911,
0.09174764156341553,
-0.06468630582094193,
0.04942422732710838,
0.0789749026298523,
0.097696952521801,
-0.008406947366893291,
0.07010164111852646,
-0.27552005648612976,
0.2783557176589966,
0.013678501360118389,
0.06311482191085815,
-0.0702686682343483,
-0.007956684567034245,
0.05028752237558365,
0.08677032589912415,
0.06043114513158798,
-0.012386336922645569,
-0.04487760365009308,
-0.23396360874176025,
-0.07560405135154724,
0.02321065217256546,
0.11294281482696533,
-0.06379324197769165,
0.11192560195922852,
-0.04183865338563919,
-0.002975957002490759,
0.05845116451382637,
-0.04309545457363129,
-0.05173318088054657,
-0.08039159327745438,
0.0034129454288631678,
0.008749188855290413,
-0.0075396085157990456,
-0.04893318563699722,
-0.12069528549909592,
-0.0725434198975563,
0.15218614041805267,
-0.031480785459280014,
-0.03433641418814659,
-0.13551218807697296,
0.0929778665304184,
0.09986288845539093,
-0.09155602753162384,
0.027659110724925995,
0.013986698351800442,
0.07454980909824371,
0.039382290095090866,
-0.05783145874738693,
0.11139766871929169,
-0.06506776064634323,
-0.2042085826396942,
-0.0641968622803688,
0.10386475920677185,
0.04504629597067833,
0.07124533504247665,
-0.022456545382738113,
0.025449031963944435,
-0.020415112376213074,
-0.08411399275064468,
0.026505300775170326,
-0.012561464682221413,
0.05334462970495224,
0.04492533951997757,
-0.06666020303964615,
0.045654296875,
-0.06331799179315567,
-0.008857492357492447,
0.15970847010612488,
0.24593165516853333,
-0.10225869715213776,
0.0229534562677145,
0.030548855662345886,
-0.06602834910154343,
-0.20636846125125885,
0.040281906723976135,
0.07423587888479233,
0.024776646867394447,
0.0392996147274971,
-0.20419985055923462,
0.08838513493537903,
0.095907062292099,
-0.010989957489073277,
0.09056863188743591,
-0.3025849163532257,
-0.12709151208400726,
0.1073194071650505,
0.12605252861976624,
0.12450920045375824,
-0.14317475259304047,
-0.019764259457588196,
-0.010369389317929745,
-0.08215457946062088,
0.09010785818099976,
-0.06284702569246292,
0.12890776991844177,
-0.02889266237616539,
0.08876053988933563,
0.029831448569893837,
-0.04491262137889862,
0.11748261004686356,
0.017399435862898827,
0.09350244700908661,
-0.04887935891747475,
-0.04633784666657448,
0.022313565015792847,
-0.04921181499958038,
0.0081315403804183,
-0.0764734297990799,
0.04630560055375099,
-0.11744267493486404,
-0.018115462735295296,
-0.08941791206598282,
0.021513627842068672,
-0.03648073971271515,
-0.06455259770154953,
-0.016483735293149948,
0.03765342757105827,
0.04625909775495529,
-0.0070994081906974316,
0.1201590895652771,
-0.014997022226452827,
0.14855149388313293,
0.10948082059621811,
0.08525358885526657,
-0.019982870668172836,
-0.021290099248290062,
0.00156973407138139,
-0.0052634733729064465,
0.055137451738119125,
-0.12583692371845245,
0.033742669969797134,
0.14913898706436157,
0.034307438880205154,
0.12844866514205933,
0.08324781060218811,
-0.0011301030172035098,
0.00870197918266058,
0.0524178147315979,
-0.1719714254140854,
-0.05923989415168762,
-0.014957034029066563,
-0.06567585468292236,
-0.13175323605537415,
0.04884384945034981,
0.11211905628442764,
-0.06834476441144943,
-0.012830058112740517,
-0.011260652914643288,
0.011558689177036285,
-0.037082646042108536,
0.2176973670721054,
0.059596315026283264,
0.05977341905236244,
-0.10424703359603882,
0.06424706429243088,
0.04978356882929802,
-0.05716651678085327,
0.007447395008057356,
0.0997227132320404,
-0.09034059196710587,
-0.036548178642988205,
0.060887355357408524,
0.15306523442268372,
-0.07600223273038864,
-0.01777099445462227,
-0.15725158154964447,
-0.1119522750377655,
0.07188520580530167,
0.17935830354690552,
0.09961584955453873,
0.02096487022936344,
-0.06636892259120941,
0.02260301075875759,
-0.12266381829977036,
0.09682681411504745,
0.05733294039964676,
0.0769181102514267,
-0.14432363212108612,
0.18738886713981628,
-0.02162599191069603,
0.04966329038143158,
-0.029918117448687553,
0.021953711286187172,
-0.10980457812547684,
0.010691234841942787,
-0.13085466623306274,
-0.034247372299432755,
-0.006939345970749855,
0.0006043798057362437,
-0.011387786827981472,
-0.07481002062559128,
-0.0479993112385273,
0.0018827127059921622,
-0.11436966061592102,
-0.025666318833827972,
0.019687844440340996,
0.03507358208298683,
-0.12077423930168152,
-0.04038889706134796,
0.03220236673951149,
-0.08045089989900589,
0.07908812910318375,
0.05944797396659851,
0.02029167115688324,
0.050141558051109314,
-0.12467695772647858,
0.013942011632025242,
0.032527633011341095,
0.00886548962444067,
0.06333908438682556,
-0.06685300916433334,
0.0022004444617778063,
-0.03450453281402588,
0.06250985711812973,
0.021605193614959717,
0.10969400405883789,
-0.11888956278562546,
0.030067941173911095,
-0.010493315756320953,
-0.06961110234260559,
-0.0650479644536972,
0.05876419320702553,
0.07402043789625168,
0.030710147693753242,
0.16310837864875793,
-0.08928047865629196,
0.05212852358818054,
-0.21827958524227142,
0.00011262119369348511,
-0.008673243224620819,
-0.12059034407138824,
-0.1313066929578781,
-0.07990743219852448,
0.08901116997003555,
-0.04752931743860245,
0.09185881912708282,
0.018581779673695564,
0.09739580005407333,
0.024750465527176857,
-0.0336892306804657,
0.033112797886133194,
0.03146826848387718,
0.18050517141819,
0.04946858063340187,
-0.04865157604217529,
0.07587865740060806,
0.052941396832466125,
0.0905272513628006,
0.14638282358646393,
0.22050167620182037,
0.1252925843000412,
-0.014161298051476479,
0.09095361828804016,
0.04152556136250496,
-0.07318338006734848,
-0.16509100794792175,
0.004805115982890129,
-0.05136844888329506,
0.08589021116495132,
-0.04557950049638748,
0.21413874626159668,
0.04665890336036682,
-0.167206808924675,
0.04787522554397583,
-0.04955203831195831,
-0.10340172052383423,
-0.12438362091779709,
-0.013362245634198189,
-0.0810261145234108,
-0.14142504334449768,
-0.0014637387357652187,
-0.11006352305412292,
0.03833537921309471,
0.1070430725812912,
0.010029719211161137,
-0.009660710580646992,
0.16120821237564087,
0.02186957374215126,
0.03827500343322754,
0.07077738642692566,
0.01226293295621872,
-0.00365520641207695,
-0.08043073117733002,
-0.09344708174467087,
-0.004330773372203112,
-0.015032714232802391,
0.031330909579992294,
-0.07359488308429718,
-0.05778754502534866,
0.03160840645432472,
-0.005757007282227278,
-0.11025238782167435,
0.014930598437786102,
0.013444327749311924,
0.07068105787038803,
0.06370586901903152,
0.009051067754626274,
0.011855201795697212,
-0.021768134087324142,
0.2453591376543045,
-0.09494245797395706,
-0.044670432806015015,
-0.10961588472127914,
0.29075732827186584,
0.03412967920303345,
0.0013501999201253057,
0.026082338765263557,
-0.07273059338331223,
-0.009712327271699905,
0.22958694398403168,
0.17701543867588043,
-0.11602586507797241,
-0.008703319355845451,
0.005419441964477301,
-0.007101657800376415,
-0.0037247007712721825,
0.10774396359920502,
0.09718601405620575,
0.037887655198574066,
-0.09969376772642136,
-0.041815031319856644,
-0.0257548950612545,
-0.031523872166872025,
-0.031076598912477493,
0.07867945730686188,
0.02758428454399109,
0.018819816410541534,
-0.05990422144532204,
0.043483275920152664,
-0.07226118445396423,
-0.1095646470785141,
0.06913253664970398,
-0.21576060354709625,
-0.17402976751327515,
-0.028186531737446785,
0.062293827533721924,
0.021847598254680634,
0.07239795476198196,
-0.023722181096673012,
-0.006693373899906874,
0.079352967441082,
-0.016941601410508156,
-0.061556991189718246,
-0.10779332369565964,
0.10494333505630493,
-0.0946192666888237,
0.19704186916351318,
-0.04746856167912483,
0.047466374933719635,
0.11663972586393356,
0.049819838255643845,
-0.0750952959060669,
0.0437769815325737,
0.05918814614415169,
-0.0951383039355278,
0.025428466498851776,
0.13922183215618134,
-0.04878917708992958,
0.08060284703969955,
0.034756384789943695,
-0.15183496475219727,
0.018608128651976585,
-0.0726482942700386,
-0.0744599923491478,
-0.030573910102248192,
-0.045697733759880066,
-0.035932496190071106,
0.14088229835033417,
0.25023528933525085,
-0.02909814938902855,
0.013158116489648819,
-0.07496234029531479,
0.0036562164314091206,
0.054802071303129196,
0.05488162860274315,
-0.06507939100265503,
-0.23882588744163513,
0.0029338125605136156,
0.07047823816537857,
-0.010928221046924591,
-0.23538093268871307,
-0.08837035298347473,
0.0047914632596075535,
-0.05356329679489136,
-0.08296243846416473,
0.09781280905008316,
0.055930234491825104,
0.043379466980695724,
-0.04375354200601578,
-0.06669451296329498,
-0.06585916876792908,
0.1641789972782135,
-0.15685336291790009,
-0.08487405627965927
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.