Skip to content

AbstractVision documentation

This folder contains the user-facing documentation for abstractvision.

Start here (new users)

1) Project overview + quickstart
2) Getting started (first image with Stable Diffusion 1.5; then klein-4B, GGUF, OpenAI-compatible HTTP, Playground) 3) Architecture (how the pieces fit together)

Quick reference

AbstractFramework ecosystem

AbstractVision is part of the AbstractFramework ecosystem and is designed to compose with:

Current implementation status (as shipped)

Public API surface: VisionManager exposes: - generate_image (text_to_image), edit_image (image_to_image) - generate_video (text_to_video), image_to_video (image_to_video) (backend-dependent) - generate_angles (multi_view_image) (API exists; no built-in backend implements it yet)

Built-in backends implement: - Images: Diffusers, stable-diffusion.cpp, OpenAI-compatible HTTP (../src/abstractvision/backends/) - Video: OpenAI-compatible HTTP only, and only when endpoints are configured (openai_compatible.py)

If you’re looking for “what can model X do?”, the single source of truth is the packaged registry: ../src/abstractvision/assets/vision_model_capabilities.json (loaded by VisionModelCapabilitiesRegistry in ../src/abstractvision/model_capabilities.py).

Internal engineering notes

docs/backlog/ is an internal log (planned work + completion reports). It is not the normative user documentation surface.

Project