Last updated:
0 purchases
overeasy 0.2.16
🥚 Overeasy
Create powerful zero-shot vision models!
Overeasy allows you to chain zero-shot vision models to create custom end-to-end pipelines for tasks like:
📦 Bounding Box Detection
🏷️ Classification
🖌️ Segmentation (Coming Soon!)
All of this can be achieved without needing to collect and annotate large training datasets.
Overeasy makes it simple to combine pre-trained zero-shot models to build powerful custom computer vision solutions.
Installation
It's as easy as
pip install overeasy
For installing extras refer to our Docs.
Key Features
🤖 Agents: Specialized tools that perform specific image processing tasks.
🧩 Workflows: Define a sequence of Agents to process images in a structured manner.
🔗 Execution Graphs: Manage and visualize the image processing pipeline.
🔎 Detections: Represent bounding boxes, segmentation, and classifications.
Documentation
For more details on types, library structure, and available models please refer to our Docs.
Example Usage
Note: If you don't have a local GPU, you can run our examples by making a copy of this Colab notebook.
Download example image
!wget https://github.com/overeasy-sh/overeasy/blob/73adbaeba51f532a7023243266da826ed1ced6ec/examples/construction.jpg?raw=true -O construction.jpg
Example workflow to identify if a person is wearing a PPE on a work site:
from overeasy import *
from overeasy.models import OwlV2
from PIL import Image
workflow = Workflow([
# Detect each head in the input image
BoundingBoxSelectAgent(classes=["person's head"], model=OwlV2()),
# Applies Non-Maximum Suppression to remove overlapping bounding boxes
NMSAgent(iou_threshold=0.5, score_threshold=0),
# Splits the input image into images of each detected head
SplitAgent(),
# Classifies the split images using CLIP
ClassificationAgent(classes=["hard hat", "no hard hat"]),
# Maps the returned class names
ClassMapAgent({"hard hat": "has ppe", "no hard hat": "no ppe"}),
# Combines results back into a BoundingBox Detection
JoinAgent()
])
image = Image.open("./construction.jpg")
result, graph = workflow.execute(image)
workflow.visualize(graph)
Diagram
Here's a diagram of this workflow. Each layer in the graph represents a step in the workflow:
The image and data attributes in each node are used together to visualize the current state of the workflow. Calling the visualize function on the workflow will spawn a Gradio instance that looks like this.
Support
If you have any questions or need assistance, please open an issue or reach out to us at [email protected].
Let's build amazing vision models together 🍳!
For personal and professional use. You cannot resell or redistribute these repositories in their original state.
There are no reviews.