File size: 2,302 Bytes
5cfc00e
e936c95
 
 
 
 
 
5cfc00e
 
c583915
26789f8
e936c95
338f580
e936c95
d6385ee
e936c95
 
 
d6385ee
c583915
d6385ee
e936c95
 
d6385ee
e936c95
d50117f
e936c95
 
d6385ee
 
e936c95
 
d6385ee
e936c95
d50117f
e936c95
 
 
 
c583915
e936c95
 
 
d6385ee
e936c95
 
d6385ee
e936c95
d6385ee
e936c95
 
 
 
 
d50117f
c583915
d6385ee
e936c95
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
---
title: README  
emoji: πŸ“š  
colorFrom: pink  
colorTo: gray  
sdk: static  
pinned: false  
---

# πŸ“š BigLAM: Machine Learning for Libraries, Archives, and Museums

**BigLAM** is a community-driven initiative to build an open ecosystem of machine learning models, datasets, and tools for **Libraries, Archives, and Museums (LAMs)**.

We aim to:

- πŸ—ƒοΈ Share machine-learning-ready datasets from LAMs via the [Hugging Face Hub](https://huggingface.co/biglam)
- πŸ€– Train and release open-source models for LAM-relevant tasks
- πŸ› οΈ Develop tools and approaches tailored to LAM use cases

---

<details>
<summary><strong>✨ Background</strong></summary>

BigLAM began as a [datasets hackathon](https://github.com/bigscience-workshop/lam) within the [BigScience 🌸](https://bigscience.huggingface.co/) project, a large-scale, open NLP collaboration.  

Our goal: make LAM datasets more discoverable and usable to support researchers, institutions, and ML practitioners working with cultural heritage data.
</details>


<details>
<summary><strong>πŸ“‚ What You'll Find</strong></summary>

The [BigLAM organization](https://huggingface.co/biglam) hosts:

- **Datasets**: image, text, and tabular data from and about libraries, archives, and museums
- **Models**: fine-tuned for tasks like:
  - Art/historical image classification
  - Document layout analysis and OCR
  - Metadata quality assessment
  - Named entity recognition in heritage texts
- **Spaces**: tools for interactive exploration and demonstration
</details>

<details>
<summary><strong>🧩 Get Involved</strong></summary>

We welcome contributions! You can:

- Use our [datasets and models](https://huggingface.co/biglam)
- Join the discussion on [GitHub](https://github.com/bigscience-workshop/lam/discussions)
- Contribute your own tools or data
- Share your work using BigLAM resources
</details>

## 🌍 Why It Matters

Cultural heritage data is often underrepresented in machine learning. BigLAM helps address this by:

- Supporting inclusive and responsible AI
- Helping institutions experiment with ML for access, discovery, and preservation
- Ensuring that ML systems reflect diverse human knowledge and expression
- Developing tools and methods that work well with the unique formats, values, and needs of LAMs