Spaces:
Running
Running
title: Starfish - Synthetic Data Generation | |
emoji: 🌟 | |
colorFrom: pink | |
colorTo: blue | |
sdk: docker | |
sdk_version: "4.36.0" | |
<p align="center"> | |
<img src="https://github.com/user-attachments/assets/744c666a-bb5c-418b-aab4-162072c0b8c8" alt="Starfish Logo" width="200"/> | |
</p> | |
<h1 align="center">Starfish</h1> | |
<h3 align="center" style="font-size: 20px; margin-bottom: 4px">Synthetic Data Generation Made Easy</h2> | |
</br> | |
<div align="center"> | |
[](https://github.com/starfishdata/starfish) [](https://x.com/starfishdata) [](https://huggingface.co/starfishdata) [](https://discord.gg/qWKmeUtb) | |
<br> | |
[](https://starfishdata.ai/) | |
[](https://deepwiki.com/starfishdata/starfish/1-overview) | |
</div> | |
## Overview | |
Starfish is a Python library that helps you build synthetic data your way. We adapt to your workflow—not the other way around. By combining structured LLM outputs with efficient parallel processing, Starfish lets you define exactly how your data should look and scale seamlessly from experiments to production. | |
⭐ Star us on GitHub if you find this project useful! | |
Key Features: | |
- **Structured Outputs**: First-class support for structured data through JSON schemas or Pydantic models. | |
- **Model Flexibility**: Use any LLM provider—local models, OpenAI, Anthropic, or your own implementation via LiteLLM. | |
- **Dynamic Prompts**: Dynamic prompts with built-in Jinja2 templates. | |
- **Easy Scaling**: Transform any function to run in parallel across thousands of inputs with a single decorator. | |
- **Resilient Pipeline**: Automatic retries, error handling, and job resumption—pause and continue your data generation anytime. | |
- **Complete Control**: Share state across your pipeline, extend functionality with custom hooks. | |
**Official Website**: [starfishdata.ai](https://starfishdata.ai/) - We offer both self-service and managed solutions. Visit our website to explore our services or contact us for more options! | |
## Installation | |
```bash | |
pip install starfish-core | |
``` | |
### Optional Dependencies | |
Starfish supports optional dependencies for specific file parsers. Install only what you need: | |
```bash | |
# Install specific parsers | |
pip install "starfish-core[pdf]" # PDF support | |
pip install "starfish-core[docx]" # Word document support | |
pip install "starfish-core[ppt]" # PowerPoint support | |
pip install "starfish-core[excel]" # Excel support | |
pip install "starfish-core[youtube]" # YouTube support | |
# Install all parser dependencies | |
pip install "starfish-core[all]" | |
``` | |
## Configuration | |
Starfish uses environment variables for configuration. We provide a `.env.template` file to help you get started quickly: | |
```bash | |
# Copy the template to .env | |
cp .env.template .env | |
# Edit with your API keys and configuration | |
nano .env # or use your preferred editor | |
``` | |
The template includes settings for API keys, model configurations, and other runtime parameters. | |
## Quick Start | |
### Structured LLM - Type-Safe Outputs from Any Model | |
```python | |
# 1. Define structured outputs with schema | |
from starfish import StructuredLLM | |
from pydantic import BaseModel | |
# Option A: Use Pydantic for type safety | |
class QnASchema(BaseModel): | |
question: str | |
answer: str | |
# Option B: Or use simple JSON schema | |
json_schema = [ | |
{'name': 'question', 'type': 'str'}, | |
{'name': 'answer', 'type': 'str'}, | |
] | |
# 2. Create a structured LLM with your preferred output format | |
qna_llm = StructuredLLM( | |
model_name="openai/gpt-4o-mini", | |
prompt="Generate facts about {{city}}", | |
output_schema=QnASchema # or json_schema | |
) | |
# 3. Get structured responses | |
response = await qna_llm.run(city="San Francisco") | |
# Access typed data | |
print(response.data) | |
# [{'question': 'What is the iconic symbol of San Francisco?', | |
# 'answer': 'The Golden Gate Bridge is the iconic symbol of San Francisco, completed in 1937.'}] | |
# Access raw API response for complete flexibility | |
print(response.raw) # Full API object with function calls, reasoning tokens, etc. | |
``` | |
### Data Factory - Scale Any Workflow with One Decorator | |
```python | |
# Turn any function into a scalable data pipeline | |
from starfish import data_factory | |
# Works with any function - simple or complex workflows | |
@data_factory(max_concurrency=50) | |
async def parallel_qna_llm(city): | |
# This could be any arbitrary complex workflow: | |
# - Pre-processing | |
# - Multiple LLM calls | |
# - Post-processing | |
# - Error handling | |
response = await qna_llm.run(city=city) | |
return response.data | |
# Process 100 cities with 50 concurrent workers - finishes in seconds | |
cities = ["San Francisco", "New York", "Tokyo", "Paris", "London"] * 20 | |
results = parallel_qna_llm.run(city=cities) | |
# dry run to test the workflow and data | |
results = parallel_qna_llm.dry_run(city=cities) | |
# resume job which pick up from where it left off. | |
results = parallel_qna_llm.resume() | |
``` | |
### Examples | |
Check out our example notebooks for detailed walkthroughs: | |
- [Structured LLM Examples](examples/structured_llm.ipynb) | |
- [Data Factory Examples](examples/data_factory.ipynb) | |
## Documentation | |
Comprehensive documentation is on the way! | |
## Contributing | |
We'd love your help making Starfish better! Whether you're fixing bugs, adding features, or improving documentation, your contributions are welcome. | |
1. Fork the repository | |
2. Create your feature branch (`git checkout -b feature/amazing-feature`) | |
3. Commit your changes (`git commit -m 'Add some amazing feature'`) | |
4. Push to the branch (`git push origin feature/amazing-feature`) | |
5. Open a Pull Request | |
Contribution guidelines coming soon! | |
## License | |
This project is licensed under the Apache License 2.0 - see the [LICENSE](LICENSE) file for details. | |
## Contact | |
If you have any questions or feedback, feel free to reach out to us at [[email protected]](mailto:[email protected]). | |
Want to discuss your use case directly? [Schedule a meeting with our team](https://calendly.com/d/crsb-ckq-fv2/chat-with-starfishdata-team). | |
## Telemetry | |
Starfish collects minimal and anonymous telemetry data to help improve the library. Participation is optional and you can opt out by setting `TELEMETRY_ENABLED=false` in your environment variables. | |
## Citation | |
If you use Starfish in your research, please consider citing us! | |
``` | |
@software{starfish, | |
author = {Wendao, John, Ayush}, | |
title = {{Starfish: A Tool for Synthetic Data Generation}}, | |
year = {2025}, | |
url = {https://github.com/starfishdata/starfish}, | |
} | |
``` | |