Spaces:
Running
Running
File size: 2,302 Bytes
5cfc00e e936c95 5cfc00e c583915 26789f8 e936c95 338f580 e936c95 d6385ee e936c95 d6385ee c583915 d6385ee e936c95 d6385ee e936c95 d50117f e936c95 d6385ee e936c95 d6385ee e936c95 d50117f e936c95 c583915 e936c95 d6385ee e936c95 d6385ee e936c95 d6385ee e936c95 d50117f c583915 d6385ee e936c95 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 |
---
title: README
emoji: π
colorFrom: pink
colorTo: gray
sdk: static
pinned: false
---
# π BigLAM: Machine Learning for Libraries, Archives, and Museums
**BigLAM** is a community-driven initiative to build an open ecosystem of machine learning models, datasets, and tools for **Libraries, Archives, and Museums (LAMs)**.
We aim to:
- ποΈ Share machine-learning-ready datasets from LAMs via the [Hugging Face Hub](https://huggingface.co/biglam)
- π€ Train and release open-source models for LAM-relevant tasks
- π οΈ Develop tools and approaches tailored to LAM use cases
---
<details>
<summary><strong>β¨ Background</strong></summary>
BigLAM began as a [datasets hackathon](https://github.com/bigscience-workshop/lam) within the [BigScience πΈ](https://bigscience.huggingface.co/) project, a large-scale, open NLP collaboration.
Our goal: make LAM datasets more discoverable and usable to support researchers, institutions, and ML practitioners working with cultural heritage data.
</details>
<details>
<summary><strong>π What You'll Find</strong></summary>
The [BigLAM organization](https://huggingface.co/biglam) hosts:
- **Datasets**: image, text, and tabular data from and about libraries, archives, and museums
- **Models**: fine-tuned for tasks like:
- Art/historical image classification
- Document layout analysis and OCR
- Metadata quality assessment
- Named entity recognition in heritage texts
- **Spaces**: tools for interactive exploration and demonstration
</details>
<details>
<summary><strong>π§© Get Involved</strong></summary>
We welcome contributions! You can:
- Use our [datasets and models](https://huggingface.co/biglam)
- Join the discussion on [GitHub](https://github.com/bigscience-workshop/lam/discussions)
- Contribute your own tools or data
- Share your work using BigLAM resources
</details>
## π Why It Matters
Cultural heritage data is often underrepresented in machine learning. BigLAM helps address this by:
- Supporting inclusive and responsible AI
- Helping institutions experiment with ML for access, discovery, and preservation
- Ensuring that ML systems reflect diverse human knowledge and expression
- Developing tools and methods that work well with the unique formats, values, and needs of LAMs
|