Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# EmbedNeural
|
| 2 |
+
|
| 3 |
+
*On-device multimodal embedding model enabling instant, private NPU-powered visual search.*
|
| 4 |
+
|
| 5 |
+
---
|
| 6 |
+
|
| 7 |
+
## Model Description
|
| 8 |
+
|
| 9 |
+
**EmbedNeural** is the world’s first multimodal embedding model purpose-built for **Qualcomm Hexagon NPU** devices. It enables **instant, private, battery-efficient** natural-language image search directly on laptops, phones, XR, and edge devices — with no cloud and no uploads.
|
| 10 |
+
|
| 11 |
+
The model continuously indexes local images using NPU acceleration, turning unorganized photo folders into a fully searchable visual database that runs entirely on-device.
|
| 12 |
+
|
| 13 |
+
---
|
| 14 |
+
|
| 15 |
+
## Key Features
|
| 16 |
+
|
| 17 |
+
### ⚡ NPU-accelerated multimodal embeddings
|
| 18 |
+
Optimized for Qualcomm NPUs to deliver sub-second search and dramatically lower power consumption.
|
| 19 |
+
|
| 20 |
+
### 🔍 Natural-language visual search
|
| 21 |
+
Query thousands of images instantly using everyday language (e.g., “green bedroom aesthetic”, “cat wearing sunglasses”).
|
| 22 |
+
|
| 23 |
+
### 🔒 100% local and private
|
| 24 |
+
All computation stays on-device. No cloud. No upload. No tracking.
|
| 25 |
+
|
| 26 |
+
### 🔋 Ultra-low power
|
| 27 |
+
Continuous background indexing uses ~10× less power than CPU/GPU methods, enabling true always-on search.
|
| 28 |
+
|
| 29 |
+
---
|
| 30 |
+
|
| 31 |
+
## Why It Matters
|
| 32 |
+
|
| 33 |
+
People save thousands of images — memes, screenshots, design inspo, photos — but struggle to find them when needed. Cloud solutions compromise privacy; CPU/GPU search drains battery.
|
| 34 |
+
|
| 35 |
+
EmbedNeural removes these tradeoffs by combining:
|
| 36 |
+
- **Instant retrieval** (~0.03s across thousands of images)
|
| 37 |
+
- **Continuous local indexing**
|
| 38 |
+
- **Zero data upload**
|
| 39 |
+
- **NPU-optimized efficiency for daily use**
|
| 40 |
+
|
| 41 |
+
This makes visual search something you can actually use **every day**, not just when plugged in.
|
| 42 |
+
|
| 43 |
+
---
|
| 44 |
+
|
| 45 |
+
## Use Cases
|
| 46 |
+
|
| 47 |
+
- **Personal image libraries:** Rediscover memes, screenshots, and old photos instantly.
|
| 48 |
+
- **Creative workflows:** Search moodboards and visual references with natural language.
|
| 49 |
+
- **Edge & embedded systems:** Efficient multimodal search for mobile, XR, IoT, and automotive.
|
| 50 |
+
|
| 51 |
+
---
|
| 52 |
+
|
| 53 |
+
## Performance Highlights
|
| 54 |
+
|
| 55 |
+
- Sub-second search even across large image libraries
|
| 56 |
+
- ~10× lower power consumption vs CPU/GPU search
|
| 57 |
+
- Stable always-on indexing without thermal or battery issues
|