|  | --- | 
					
						
						|  | tags: | 
					
						
						|  | - bert | 
					
						
						|  |  | 
					
						
						|  | --- | 
					
						
						|  | # Model Card for bert-small-mm_retrieval-table_encoder | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | # Model Details | 
					
						
						|  |  | 
					
						
						|  | ## Model Description | 
					
						
						|  |  | 
					
						
						|  | - **Developed by:** deepset | 
					
						
						|  | - **Shared by [Optional]:** More information needed | 
					
						
						|  | - **Model type:** More information needed | 
					
						
						|  | - **Language(s) (NLP):** More information needed | 
					
						
						|  | - **License:** More information needed | 
					
						
						|  | - **Related Models:** | 
					
						
						|  | - **Parent Model:** More information needed | 
					
						
						|  | - **Resources for more information:** | 
					
						
						|  | - [Associated Paper](https://arxiv.org/abs/1908.08962) | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | # Uses | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | ## Direct Use | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ## Downstream Use [Optional] | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ## Out-of-Scope Use | 
					
						
						|  |  | 
					
						
						|  | The model should not be used to intentionally create hostile or alienating environments for people. | 
					
						
						|  |  | 
					
						
						|  | # Bias, Risks, and Limitations | 
					
						
						|  |  | 
					
						
						|  | Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups. | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | ## Recommendations | 
					
						
						|  |  | 
					
						
						|  | Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | # Training Details | 
					
						
						|  |  | 
					
						
						|  | ## Training Data | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ## Training Procedure | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | ### Preprocessing | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ### Speeds, Sizes, Times | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | # Evaluation | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | ## Testing Data, Factors & Metrics | 
					
						
						|  |  | 
					
						
						|  | ### Testing Data | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ### Factors | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | ### Metrics | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  | ## Results | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | # Model Examination | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | # Environmental Impact | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). | 
					
						
						|  |  | 
					
						
						|  | - **Hardware Type:** More information needed | 
					
						
						|  | - **Hours used:** More information needed | 
					
						
						|  | - **Cloud Provider:** More information needed | 
					
						
						|  | - **Compute Region:** More information needed | 
					
						
						|  | - **Carbon Emitted:** More information needed | 
					
						
						|  |  | 
					
						
						|  | # Technical Specifications [optional] | 
					
						
						|  |  | 
					
						
						|  | ## Model Architecture and Objective | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ## Compute Infrastructure | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ### Hardware | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | ### Software | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | # Citation | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | **BibTeX:** | 
					
						
						|  | ``` | 
					
						
						|  | @misc{bhargava2021generalization, | 
					
						
						|  | title={Generalization in NLI: Ways (Not) To Go Beyond Simple Heuristics}, | 
					
						
						|  | author={Prajjwal Bhargava and Aleksandr Drozd and Anna Rogers}, | 
					
						
						|  | year={2021}, | 
					
						
						|  | eprint={2110.01518}, | 
					
						
						|  | archivePrefix={arXiv}, | 
					
						
						|  | primaryClass={cs.CL} | 
					
						
						|  | } | 
					
						
						|  |  | 
					
						
						|  | @article{DBLP:journals/corr/abs-1908-08962, | 
					
						
						|  | author    = {Iulia Turc and | 
					
						
						|  | Ming{-}Wei Chang and | 
					
						
						|  | Kenton Lee and | 
					
						
						|  | Kristina Toutanova}, | 
					
						
						|  | title     = {Well-Read Students Learn Better: The Impact of Student Initialization | 
					
						
						|  | on Knowledge Distillation}, | 
					
						
						|  | journal   = {CoRR}, | 
					
						
						|  | volume    = {abs/1908.08962}, | 
					
						
						|  | year      = {2019}, | 
					
						
						|  | url       = {http://arxiv.org/abs/1908.08962}, | 
					
						
						|  | eprinttype = {arXiv}, | 
					
						
						|  | eprint    = {1908.08962}, | 
					
						
						|  | timestamp = {Thu, 29 Aug 2019 16:32:34 +0200}, | 
					
						
						|  | biburl    = {https://dblp.org/rec/journals/corr/abs-1908-08962.bib}, | 
					
						
						|  | bibsource = {dblp computer science bibliography, https://dblp.org} | 
					
						
						|  | } | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | ``` | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | # Glossary [optional] | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | # More Information [optional] | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | # Model Card Authors [optional] | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | Deepset in collaboration with Ezi Ozoani and the Hugging Face team | 
					
						
						|  |  | 
					
						
						|  | # Model Card Contact | 
					
						
						|  |  | 
					
						
						|  | More information needed | 
					
						
						|  |  | 
					
						
						|  | # How to Get Started with the Model | 
					
						
						|  |  | 
					
						
						|  | Use the code below to get started with the model. | 
					
						
						|  |  | 
					
						
						|  | <details> | 
					
						
						|  | <summary> Click to expand </summary> | 
					
						
						|  |  | 
					
						
						|  | ```python | 
					
						
						|  | from transformers import AutoTokenizer, DPRContextEncoder | 
					
						
						|  |  | 
					
						
						|  | tokenizer = AutoTokenizer.from_pretrained("deepset/bert-small-mm_retrieval-table_encoder") | 
					
						
						|  |  | 
					
						
						|  | model = DPRContextEncoder.from_pretrained("deepset/bert-small-mm_retrieval-table_encoder") | 
					
						
						|  |  | 
					
						
						|  | ``` | 
					
						
						|  | </details> | 
					
						
						|  |  |