GenAIDevTOProd commited on
Commit
edcedb1
·
verified ·
1 Parent(s): 76ef4fb

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,12 +1,3 @@
1
- ---
2
- title: PromptGuru
3
- emoji: 🧠
4
- colorFrom: indigo
5
- colorTo: purple
6
- sdk: static
7
- pinned: false
8
- ---
9
-
10
  # PromptGuru
11
 
12
  **Modular prompt engineering library** for BERT, Mistral, LLaMA, and FLAN-T5 using YAML templates.
 
 
 
 
 
 
 
 
 
 
1
  # PromptGuru
2
 
3
  **Modular prompt engineering library** for BERT, Mistral, LLaMA, and FLAN-T5 using YAML templates.
dist/promptguru-0.1.0-py3-none-any.whl ADDED
Binary file (4.37 kB). View file
 
dist/promptguru-0.1.0.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b49f252187399b6e1be5864a5df37a2c9f71427b41010425fa45ad6ad0a856a
3
+ size 3258
promptguru.egg-info/PKG-INFO ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metadata-Version: 2.4
2
+ Name: promptguru
3
+ Version: 0.1.0
4
+ Summary: Modular prompt engineering library
5
+ Author-email: Naga Adithya Kaushik <[email protected]>
6
+ License: Apache-2.0
7
+ Project-URL: Homepage, https://huggingface.co/spaces/GenAIDevTOProd/PromptGuru
8
+ Keywords: prompts,nlp,llm,huggingface,templates
9
+ Requires-Python: >=3.8
10
+ Description-Content-Type: text/markdown
11
+ Requires-Dist: PyYAML>=6.0
12
+
13
+ # PromptGuru
14
+
15
+ **Modular prompt engineering library** for BERT, Mistral, LLaMA, and FLAN-T5 using YAML templates.
16
+ Includes modes like **ELI5**, **DevMode**, **Refine**, **Classification**, and **QA**.
17
+
18
+ ## Why
19
+ - Lightweight and framework-agnostic
20
+ - YAML-first: edit prompts without changing code
21
+ - Consistent modes across multiple model families
22
+
23
+ ## Install (Local Dev)
24
+ ```bash
25
+ pip install PyYAML
26
+ ```
27
+ > For now, clone or copy this repo. PyPI packaging steps are included below.
28
+
29
+ ## Usage
30
+ ```python
31
+ from promptguru.engine import PromptEngine
32
+
33
+ engine = PromptEngine(model_type="mistral", mode="eli5")
34
+ prompt = engine.generate_prompt("What is quantum entanglement?")
35
+ print(prompt)
36
+ ```
37
+
38
+ ## Templates
39
+ Templates live in `promptguru/templates/`:
40
+ - `bert.yaml` → `classification`, `fill_mask`, `qa`
41
+ - `mistral.yaml` → `eli5`, `devmode`, `refine`
42
+ - `llama.yaml` → `eli5`, `devmode`, `refine`
43
+ - `flan_t5.yaml` → `eli5`, `devmode`, `explain_and_tag`
44
+
45
+ ## Roadmap
46
+ - Add inference adapters (HF Inference API, OpenRouter) behind a common interface
47
+ - Add more modes (contrastive QA, chain-of-thought, safety/risk tags)
48
+
49
+ ## License
50
+ Apache 2.0
promptguru.egg-info/SOURCES.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ README.md
2
+ pyproject.toml
3
+ promptguru/__init__.py
4
+ promptguru/engine.py
5
+ promptguru.egg-info/PKG-INFO
6
+ promptguru.egg-info/SOURCES.txt
7
+ promptguru.egg-info/dependency_links.txt
8
+ promptguru.egg-info/requires.txt
9
+ promptguru.egg-info/top_level.txt
10
+ promptguru/templates/bert.yaml
11
+ promptguru/templates/flan_t5.yaml
12
+ promptguru/templates/llama.yaml
13
+ promptguru/templates/mistral.yaml
promptguru.egg-info/dependency_links.txt ADDED
@@ -0,0 +1 @@
 
 
1
+
promptguru.egg-info/requires.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ PyYAML>=6.0
promptguru.egg-info/top_level.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ promptguru
promptguru/__pycache__/__init__.cpython-311.pyc ADDED
Binary file (353 Bytes). View file
 
promptguru/__pycache__/engine.cpython-311.pyc ADDED
Binary file (2.67 kB). View file