You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+25-20Lines changed: 25 additions & 20 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -58,8 +58,10 @@ Each skill is a self-contained module with its own model, parameters, and [commu
58
58
| Category | Skill | What It Does | Status |
59
59
|----------|-------|--------------|:------:|
60
60
|**Detection**|[`yolo-detection-2026`](skills/detection/yolo-detection-2026/)| Real-time 80+ class detection — auto-accelerated via TensorRT / CoreML / OpenVINO / ONNX | ✅|
61
-
||[`yolo-detection-2026-coral-tpu`](skills/detection/yolo-detection-2026-coral-tpu/)| Google Coral Edge TPU — ~4ms inference via USB accelerator ([Docker-based](#detection--segmentation-skills)) | 🧪 |
62
-
||[`yolo-detection-2026-openvino`](skills/detection/yolo-detection-2026-openvino/)| Intel NCS2 USB / Intel GPU / CPU — multi-device via OpenVINO ([Docker-based](#detection--segmentation-skills)) | 🧪 |
61
+
||[`yolo-detection-2026-coral-tpu`](skills/detection/yolo-detection-2026-coral-tpu/)| Google Coral Edge TPU — ~4ms inference via USB accelerator ([LiteRT](#detection--segmentation-skills)) | ✅ |
62
+
||[`yolo-detection-2026-openvino`](skills/detection/yolo-detection-2026-openvino/)| Intel NCS2 USB / Intel GPU / CPU — multi-device via OpenVINO ([architecture](#detection--segmentation-skills)) | 🧪 |
63
+
||`face-detection-recognition`| Face detection & recognition — identify known faces from camera feeds | 📐 |
64
+
||`license-plate-recognition`| License plate detection & recognition — read plate numbers from camera feeds | 📐 |
|**Privacy**|[`depth-estimation`](skills/transformation/depth-estimation/)|[Real-time depth-map privacy transform](#-privacy--depth-map-anonymization) — anonymize camera feeds while preserving activity | ✅ |
65
67
|**Segmentation**|[`sam2-segmentation`](skills/segmentation/sam2-segmentation/)| Interactive click-to-segment with Segment Anything 2 — pixel-perfect masks, point/box prompts, video tracking | ✅ |
@@ -74,38 +76,41 @@ Each skill is a self-contained module with its own model, parameters, and [commu
74
76
75
77
### Detection & Segmentation Skills
76
78
77
-
Detection and segmentation skills process visual data from camera feeds — detecting objects, segmenting regions, or analyzing scenes. All skills use the same **JSONL stdin/stdout protocol**: Aegis writes a frame to a shared volume, sends a `frame` event on stdin, and reads `detections` from stdout. This means every detection skill — whether running natively or inside Docker — is interchangeable from Aegis's perspective.
79
+
Detection and segmentation skills process visual data from camera feeds — detecting objects, segmenting regions, or analyzing scenes. All skills use the same **JSONL stdin/stdout protocol**: Aegis writes a frame to a shared volume, sends a `frame` event on stdin, and reads `detections` from stdout. Every detection skill is interchangeable from Aegis's perspective.
78
80
79
81
```mermaid
80
82
graph TB
81
83
CAM["📷 Camera Feed"] --> GOV["Frame Governor (5 FPS)"]
-**Native skills**run directly on the host — [`env_config.py`](skills/lib/env_config.py) auto-detects the GPU and converts models to the fastest format (TensorRT, CoreML, OpenVINO IR, ONNX)
108
-
-**Docker skills**wrap hardware-specific runtimes in a container — cross-platform USB/device access without native driver installation
112
+
-**Unified protocol**— each skill creates its own Python venv or Docker container, but Aegis sees the same JSONL interface regardless of backend
113
+
-**Coral TPU**uses [ai-edge-litert](https://pypi.org/project/ai-edge-litert/) (LiteRT) with the `libedgetpu` delegate — supports Python 3.9–3.13 on Linux, macOS, and Windows
109
114
-**Same output** — Aegis sees identical JSONL from all skills, so detection overlays, alerts, and forensic analysis work with any backend
110
115
111
116
#### LLM-Assisted Skill Installation
@@ -114,7 +119,7 @@ Skills are installed by an **autonomous LLM deployment agent** — not by brittl
114
119
115
120
1.**Probe** — reads `SKILL.md`, `requirements.txt`, and `package.json` to understand what the skill needs
116
121
2.**Detect hardware** — checks for NVIDIA (CUDA), AMD (ROCm), Apple Silicon (MPS), Intel (OpenVINO), or CPU-only
117
-
3.**Install** — runs the right commands (`pip install`, `npm install`, `docker build`) with the correct backend-specific dependencies
122
+
3.**Install** — runs the right commands (`pip install`, `npm install`, system packages) with the correct backend-specific dependencies
118
123
4.**Verify** — runs a smoke test to confirm the skill loads before marking it complete
119
124
5.**Determine launch command** — figures out the exact `run_command` to start the skill and saves it to the registry
- cd $TRAVIS_BUILD_DIR/docker/build && docker-compose -f docker-compose-arm64v8.yml push
28
+
cache:
29
+
directories:
30
+
- /home/travis/docker/
31
+
before_install:
32
+
- if [ -f ${DOCKER_CACHE_FILE} ]; then gunzip -c ${DOCKER_CACHE_FILE} | docker load; fi
33
+
before_cache:
34
+
- if [[ ${TRAVIS_BRANCH} == "master" ]] && [[ ${TRAVIS_PULL_REQUEST} == "false" ]]; then docker save $(docker images -a -q) | gzip > ${DOCKER_CACHE_FILE}; fi
Thank you for your interest in contributing to DeepCamera! This project is evolving into an open-source AI skill platform for [SharpAI Aegis](https://sharpai.org).
4
+
5
+
## How to Contribute
6
+
7
+
### 🛠️ Build a New Skill
8
+
9
+
The best way to contribute is by building a new skill. Each skill is a self-contained folder under `skills/` with:
10
+
11
+
1.**`SKILL.md`** — declares parameters (rendered as UI in Aegis) and capabilities
12
+
2.**`requirements.txt`** — Python dependencies
13
+
3.**`scripts/`** — entry point using JSON-lines stdin/stdout protocol
14
+
15
+
See [`skills/detection/yolo-detection-2026/`](skills/detection/yolo-detection-2026/) for a complete reference implementation.
16
+
17
+
### 📋 Skill Ideas We Need
18
+
19
+
- Camera providers: Eufy, Reolink, Tapo, Ring
20
+
- Messaging channels: Matrix, LINE, Signal
21
+
- Automation triggers: MQTT, webhooks
22
+
- AI models: VLM scene analysis, SAM2 segmentation, depth estimation
23
+
24
+
### 🐛 Report Issues
25
+
26
+
- Use [GitHub Issues](https://github.com/SharpAI/DeepCamera/issues)
27
+
- Include your platform, Python version, and steps to reproduce
28
+
29
+
### 📝 Improve Documentation
30
+
31
+
- Fix typos, improve clarity, add examples
32
+
- Add platform-specific setup guides under `docs/`
0 commit comments