Reverse Dependencies of open-clip-torch
The following projects have a declared dependency on open-clip-torch:
- aesthetic-predictor — A python package of aesthetic quality of pictures a.k.a `aesthic-predictor`.
- all-clip — Load any clip model with a standardized interface
- anomalib — anomalib - Anomaly Detection Library
- auto1111sdk — SDK for Automatic 1111.
- autodistill-bioclip — BioCLIP model for use with Autodistill
- autodistill-metaclip — MetaCLIP base model for use with Autodistill.
- autodistill-remote-clip — Remote CLIP model for use with Autodistill
- beprepared — no summary
- chuchichaestli — Where you find all the state-of-the-art cooking utensils (salt, pepper, gradient descent... the usual).
- classifier-free-guidance-pytorch — Classifier Free Guidance - Pytorch
- CLIP-API-service — Build AI applications with any CLIP models - embed image and sentences, object recognition, visual reasoning, image classification and reverse image search
- clip-benchmark — CLIP-like models benchmarks on various datasets
- clip-inference — clip 数据推理
- clip-interrogator — Generate a prompt from an image
- clip-retrieval — Easily computing clip embeddings and building a clip retrieval system with them
- clip-video-encode — Easily compute clip embeddings from video frames
- clorps — CLORPS: A module for CLIP, LPIPS, and ORB based image similarity.
- comfylib — Stable Diffusion library, based on the implementation in ComfyUI
- comfyui — An installable version of ComfyUI
- controlnet — ControlNet - Adding Conditional Control to Text-to-Image Diffusion Models
- dalle2-pytorch — DALL-E 2
- datachain — Wrangle unstructured AI data at scale
- diffuzers — diffuzers
- dreamai-gen — no summary
- dreamsim — DreamSim similarity metric
- dvcx — DVCx
- exordium — Collection of utility tools and deep learning methods.
- finetuner-commons — The finetuner-commons package provides common functionality between core and client.
- gem-torch — GEM
- gen1 — Text to Video synthesis
- generate-od — A tool to generate origin-destination matrix for any given area.
- genQC — Generating quantum circuits with diffusion models
- gigagan-pytorch — GigaGAN - Pytorch
- hjxdl — A collection of functions for Jupyter notebooks
- horde-engine — A wrapper around ComfyUI to allow use by AI Horde.
- hordelib — A thin wrapper around ComfyUI to allow use by AI Horde.
- hspylib-askai — HomeSetup - AskAI
- imaginarium — no summary
- imscore — A small example package
- lancedb — lancedb
- legrad-torch — LeGrad
- libcom — Image Composition Toolbox
- llm-lens — llm-lens is a Python package for CV as NLP, where you can run very descriptive image modules on images, and then pass those descriptions to a Large Language Model (LLM) to reason about those images.
- mamba-clip — A package for training mamba vision model and text encoder using CLIP
- mantis-vl — Official Codes for of "MANTIS: Interleaved Multi-Image Instruction Tuning"
- maskinversion-torch — MaskInversion
- mmagic — OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox
- mmgpt — An open-source framework for multi-modality instruction fine-tuning
- model-sketch-book — A package for sketching ML models
- modelscope — ModelScope: bring the notion of Model-as-a-Service to life.
- naifu — naifu is designed for training generative models with various configurations and features.
- nataili — Nataili: Multimodal AI Python Library
- nemo-toolkit — NeMo - a toolkit for Conversational AI
- neurosis — A neural network trainer (for weebs)
- open-clip-torch — Open reproduction of consastive language-image pretraining (CLIP) and related.
- open-flamingo — An open-source framework for training large multimodal models
- open-gpts — An open-source implementation of large-scale language model (LLM).
- openclip-service — CLIP Service Network Application - service part
- opensr-test — A comprehensive benchmark for real-world Sentinel-2 imagery super-resolution
- optimum-intel — Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality.
- otter-ai — Otter: A Multi-Modal Model with In-Context Instruction Tuning
- paella — paella - Fast Text-Conditional Discrete Denoising on Vector-Quantized Latent Spaces
- paintmind — no summary
- perceptor — Modular image generation library
- perfusion-pytorch — Perfusion - Pytorch
- pybioclip — Python package that simplifies using the BioCLIP foundation model.
- rclip — AI-Powered Command-Line Photo Search Tool
- rshf — RS pretrained models in huggingface style
- scepter — no summary
- smart-reid — With no prior knowledge of machine learning or device-specific deployment, you can deploy a computer vision model to a range of devices and environments using Roboflow Inference CLI.
- softpandas — An easy to use semantic (soft) querying on pandas dataframes.
- sparseml — Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
- sparseml-nightly — Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
- stable-diffusion-sdkit — High-Resolution Image Synthesis with Latent Diffusion Models. This is a wrapper around the original repo, to allow installing via pip.
- stablefusion — StableFusion
- stuned — Utility code from STAI (https://scalabletrustworthyai.github.io/)
- superlinked — The Superlinked vector computing library
- t2iadapter — T2I-Adapter
- t2v-metrics — Evaluating Text-to-Visual Generation with Image-to-Text Generation.
- tglcourse — work-in-progress course
- thingsvision — Extracting image features from state-of-the-art neural networks for Computer Vision made easy
- tinydata — A cli tool for building computer vision datasets.
- tooncraftersimple — Simple Tooncrafter Implementation
- tx-extension-clip — Python Library for Threat Exchange CLIP Extension
- unibench — This repository is designed to simplify the evaluation process of vision-language models. It provides a comprehensive set of tools and scripts for evaluating VLM models and benchmarks.
- v-clip-server — Embed images and sentences into fixed-length vectors via CLIP
- video_sampler — Video Sampler -- sample frames from a video file
- zerommt — An open-source framework for zero-shot multimodal machine translation inference
1