YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

EmbedNeural

On-device multimodal embedding model enabling instant, private NPU-powered visual search.

Model Description

EmbedNeural is the world’s first multimodal embedding model purpose-built for Qualcomm Hexagon NPU devices. It enables instant, private, battery-efficient natural-language image search directly on laptops, phones, XR, and edge devices β€” with no cloud and no uploads.

The model continuously indexes local images using NPU acceleration, turning unorganized photo folders into a fully searchable visual database that runs entirely on-device.


Key Features

⚑ NPU-accelerated multimodal embeddings

Optimized for Qualcomm NPUs to deliver sub-second search and dramatically lower power consumption.

πŸ” Natural-language visual search

Query thousands of images instantly using everyday language (e.g., β€œgreen bedroom aesthetic”, β€œcat wearing sunglasses”).

πŸ”’ 100% local and private

All computation stays on-device. No cloud. No upload. No tracking.

πŸ”‹ Ultra-low power

Continuous background indexing uses ~10Γ— less power than CPU/GPU methods, enabling true always-on search.


Why It Matters

People save thousands of images β€” memes, screenshots, design inspo, photos β€” but struggle to find them when needed. Cloud solutions compromise privacy; CPU/GPU search drains battery.

EmbedNeural removes these tradeoffs by combining:

  • Instant retrieval (~0.03s across thousands of images)
  • Continuous local indexing
  • Zero data upload
  • NPU-optimized efficiency for daily use

This makes visual search something you can actually use every day, not just when plugged in.


Use Cases

  • Personal image libraries: Rediscover memes, screenshots, and old photos instantly.
  • Creative workflows: Search moodboards and visual references with natural language.
  • Edge & embedded systems: Efficient multimodal search for mobile, XR, IoT, and automotive.

Performance Highlights

  • Sub-second search even across large image libraries
  • ~10Γ— lower power consumption vs CPU/GPU search
  • Stable always-on indexing without thermal or battery issues

License

This model is released under the Creative Commons Attribution–NonCommercial 4.0 (CC BY-NC 4.0) license.
Non-commercial use, modification, and redistribution are permitted with attribution.
For commercial licensing, please contact dev@nexa.ai.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Collection including NexaAI/embedneural-npu-mobile