TOP AI Developers by monthly star count
TOP AI Organization Account by AI repo star count
Top AI Project by Category star count
Top Growing Speed list by the speed of gaining stars
Top List of who create influential repos with little people known
Rankings | Organization Account | Related Project | Project intro | Star count |
---|---|---|---|---|
1 | nexa-sdk | Nexa SDK is a comprehensive toolkit for supporting GGML and ONNX models. It supports text generation, image generation, vision-language models (VLM), Audio Language Model, auto-speech-recognition (ASR), and text-to-speech (TTS) capabilities. | 5.2K | |
2 | MGM | Official repo for "Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models" | 3.2K | |
3 | cambrian | Cambrian-1 is a family of multimodal LLMs with a vision-centric design. | 1.8K | |
4 | Local-File-Organizer | An AI-powered file management tool that ensures privacy by organizing local texts, images. Using Llama3.2 3B and Llava v1.6 models with the Nexa SDK, it intuitively scans, restructures, and organizes files for quick, seamless access and easy retrieval. | 1.7K | |
5 | ShareGPT4Video | [NeurIPS 2024] An official implementation of ShareGPT4Video: Improving Video Understanding and Generation with Better Captions | 1.3K | |
6 | colpali | The code used to train and run inference with the ColPali architecture. | 1.2K | |
7 | comfyui_LLM_party | LLM Agent Framework in ComfyUI includes Omost,GPT-sovits, ChatTTS,GOT-OCR2.0, and FLUX prompt nodes,access to Feishu,discord,and adapts to all llms with similar openai / aisuite interfaces, such as o1,ollama, gemini, grok, qwen, GLM, deepseek, moonshot,doubao. Adapted to local llms, vlm, gguf such as llama-3.2, Linkage graphRAG / RAG | 1.1K | |
8 | LLaVA-pp | 🔥🔥 LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3) | 812 | |
9 | mlx-vlm | MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX. | 663 | |
10 | Groma | [ECCV2024] Grounded Multimodal Large Language Model with Localized Visual Tokenization | 568 | |
11 | Awesome-Robotics-3D | A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites | 559 | |
12 | EAGLE | EAGLE: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders | 541 | |
13 | Ovis | A novel Multimodal Large Language Model (MLLM) architecture, designed to structurally align visual and textual embeddings. | 539 | |
14 | llama-assistant | AI-powered assistant to help you with your daily tasks, powered by Llama 3.2. It can recognize your voice, process natural language, and perform various actions based on your commands: summarizing text, rephasing sentences, answering questions, writing emails, and more. | 423 | |
15 | VisRAG | Parsing-free RAG supported by VLMs | 421 | |
16 | meme_search | Index your memes by their content and text, making them easily retrievable for your meme warfare pleasures. Find funny fast. | 409 | |
17 | Awesome-Jailbreak-on-LLMs | Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, datasets, evaluations, and analyses. | 409 | |
18 | Open-LLaVA-NeXT | An open-source implementation for training LLaVA-NeXT. | 398 | |
19 | minimind-v | 「大模型」3小时从0训练27M参数的视觉多模态VLM,个人显卡即可推理训练! | 378 | |
20 | ai-devices | AI Device Template Featuring Whisper, TTS, Groq, Llama3, OpenAI and more | 281 | |
21 | RLAIF-V | RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness | 246 | |
22 | Phi-3-Vision-MLX | Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon | 237 | |
23 | EVE | [NeurIPS'24 Spotlight] EVE: Encoder-Free Vision-Language Models | 233 | |
24 | TokenPacker | The code for "TokenPacker: Efficient Visual Projector for Multimodal LLM". | 215 | |
25 | Mantis | Official code for Paper "Mantis: Multi-Image Instruction Tuning" (TMLR2024) | 189 | |
26 | rai | RAI is a multi-vendor agent framework for robotics, utilizing Langchain and ROS 2 tools to perform complex actions, defined scenarios, free interface execution, log summaries, voice interaction and more. | 178 | |
27 | seemore | From scratch implementation of a vision language model in pure PyTorch | 164 | |
28 | embodied-agents | Seamlessly integrate state-of-the-art transformer models into robotics stacks | 164 | |
29 | LLaRA | LLaRA: Large Language and Robotics Assistant | 156 | |
30 | AUITestAgent | AUITestAgent is the first automatic, natural language-driven GUI testing tool for mobile apps, capable of fully automating the entire process of GUI interaction and function verification. | 151 | |
31 | PsyDI | PsyDI: Towards a Personalized and Progressively In-depth Chatbot for Psychological Measurements. (e.g. MBTI Measurement Agent) | 151 | |
32 | image-textualization | Image Textualization: An Automatic Framework for Generating Rich and Detailed Image Descriptions (NeurIPS 2024) | 145 | |
33 | joycaption | JoyCaption is an image captioning Visual Language Model (VLM) being built from the ground up as a free, open, and uncensored model for the community to use in training Diffusion models. | 144 | |
34 | Emotion-LLaMA | Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction Tuning | 121 | |
35 | Spider2-V | [NeurIPS 2024] Spider2-V: How Far Are Multimodal Agents From Automating Data Science and Engineering Workflows? | 109 | |
36 | MMTrustEval | A toolbox for benchmarking trustworthiness of multimodal large language models (MultiTrust, NeurIPS 2024 Track Datasets and Benchmarks) | 108 | |
37 | MM-NIAH | [NeurIPS 2024] Needle In A Multimodal Haystack (MM-NIAH): A comprehensive benchmark designed to systematically evaluate the capability of existing MLLMs to comprehend long multimodal documents. | 102 | |
38 | Surveillance_Video_Summarizer | VLM driven tool that processes surveillance videos, extracts frames, and generates insightful annotations using a fine-tuned Florence-2 Vision-Language Model. Includes a Gradio-based interface for querying and analyzing video footage. | 92 | |
39 | eureka-ml-insights | A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings. | 90 | |
40 | LLaVA-MORE | LLaVA-MORE: Enhancing Visual Instruction Tuning with LLaMA 3.1 | 86 | |
41 | Llama3.2-Vision-Finetune | An open-source implementaion for fine-tuning Llama3.2-Vision series by Meta. | 85 | |
42 | Modality-Integration-Rate | The official code of the paper "Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate". | 85 | |
43 | matryoshka-mm | Matryoshka Multimodal Models | 84 | |
44 | Mini-LLaVA | A minimal implementation of LLaVA-style VLM with interleaved image & text & video processing ability. | 84 | |
45 | VoCo-LLaMA | VoCo-LLaMA: This repo is the official implementation of "VoCo-LLaMA: Towards Vision Compression with Large Language Models". | 83 | |
46 | VLM-Visualizer | Visualizing the attention of vision-language models | 79 | |
47 | BALROG | Benchmarking Agentic LLM and VLM Reasoning On Games | 77 | |
48 | CharXiv | [NeurIPS 2024] CharXiv: Charting Gaps in Realistic Chart Understanding in Multimodal LLMs | 75 | |
49 | Know-Your-Neighbors | [CVPR 2024] 🏡Know Your Neighbors: Improving Single-View Reconstruction via Spatial Vision-Language Reasoning | 69 | |
50 | YoLLaVA | 🌋👵🏻 Yo'LLaVA: Your Personalized Language and Vision Assistant | 69 | |
51 | VLM-Grounder | [CoRL 2024] VLM-Grounder: A VLM Agent for Zero-Shot 3D Visual Grounding | 69 | |
52 | SpeechLLM | This repository contains the training, inference, evaluation code for SpeechLLM models and details about the model releases on huggingface. | 61 | |
53 | STIC | Enhancing Large Vision Language Models with Self-Training on Image Comprehension. | 59 | |
54 | Elysium | [ECCV 2024] Elysium: Exploring Object-level Perception in Videos via MLLM | 58 | |
55 | CARES | [NeurIPS'24 & ICMLW'24] CARES: A Comprehensive Benchmark of Trustworthiness in Medical Vision Language Models | 56 | |
56 | SparseVLMs | Official implementation of paper "SparseVLM: Visual Token Sparsification for Efficient Vision-Language Model Inference" proposed by Peking University and UC Berkeley. | 56 | |
57 | Chinese-LLaVA-Med | 中文医学多模态大模型 Large Chinese Language-and-Vision Assistant for BioMedicine | 55 | |
58 | usls | A Rust library integrated with ONNXRuntime, providing a collection of Computer Vison and Vision-Language models. | 51 | |
59 | FreeVA | FreeVA: Offline MLLM as Training-Free Video Assistant | 49 | |
60 | KDPL | [ECCV 2024] - Improving Zero-shot Generalization of Learned Prompts via Unsupervised Knowledge Distillation | 48 | |
61 | VisualWebBench | Evaluation framework for paper "VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?" | 47 | |
62 | WCA | [ICML 2024] "Visual-Text Cross Alignment: Refining the Similarity Score in Vision-Language Models" | 43 | |
63 | VLM-Captioning-Tools | Python scripts to use for captioning images with VLMs | 34 | |
64 | MMInstruct | The official implementation of the paper "MMInstruct: A High-Quality Multi-Modal Instruction Tuning Dataset with Extensive Diversity". The MMInstruct dataset includes 973K instructions from 24 domains and four instruction types. | 34 | |
65 | AAPL | AAPL: Adding Attributes to Prompt Learning for Vision-Language Models (CVPRw 2024) | 32 | |
66 | LLaVA-UHD-Better | A bug-free and improved implementation of LLaVA-UHD, based on the code from the official repo | 32 | |
67 | ReachQA | Code & Dataset for Paper: "Distill Visual Chart Reasoning Ability from LLMs to MLLMs" | 32 | |
68 | LLM4VPR | Can multimodal LLM help visual place recognition? | 31 | |
69 | CompBench | CompBench evaluates the comparative reasoning of multimodal large language models (MLLMs) with 40K image pairs and questions across 8 dimensions of relative comparison: visual attribute, existence, state, emotion, temporality, spatiality, quantity, and quality. CompBench covers diverse visual domains, including animals, fashion, sports, and scenes. | 31 | |
70 | GMAI-MMBench | GMAI-MMBench: A Comprehensive Multimodal Evaluation Benchmark Towards General Medical AI. | 31 | |
71 | ConBench | [NeurIPS'24] Official implementation of paper "Unveiling the Tapestry of Consistency in Large Vision-Language Models". | 30 | |
72 | ThinkGrasp | [CoRL2024] ThinkGrasp: A Vision-Language System for Strategic Part Grasping in Clutter. https://arxiv.org/abs/2407.11298 | 30 | |
73 | Situation3D | [CVPR 2024] Situational Awareness Matters in 3D Vision Language Reasoning | 26 | |
74 | Jailbreak-In-Pieces | [ICLR 2024 Spotlight 🔥 ] - [ Best Paper Award SoCal NLP 2023 🏆] - Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models | 26 | |
75 | Vista | This is the official repository for Vista dataset - A Vietnamese multimodal dataset contains more than 700,000 samples of conversations and images | 24 | |
76 | Kling-AI-Webui | Kling AI, Make Imagination Alive. This is a revolutionary text-to-video model like Sora. Kling AI WebUI is the open source project to integrate Kling AI Video Generation Model. | 24 | |
77 | RET-CLIP | RET-CLIP: A Retinal Image Foundation Model Pre-trained with Clinical Diagnostic Reports | 23 | |
78 | Emma-X | Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning | 23 | |
79 | foundational_fsod | This repository contains the implementation for the paper "Revisiting Few Shot Object Detection with Vision-Language Models" | 22 | |
80 | VISTA_Evaluation_FineTuning | Evaluation code and datasets for the ACL 2024 paper, VISTA: Visualized Text Embedding for Universal Multi-Modal Retrieval. The original code and model can be accessed at FlagEmbedding. | 21 | |
81 | HOI-Ref | Code implementation for paper titled "HOI-Ref: Hand-Object Interaction Referral in Egocentric Vision" | 20 | |
82 | Vistral-V | Vistral-V: Visual Instruction Tuning for Vistral - Vietnamese Large Vision-Language Model. | 19 | |
83 | Llava_Qwen2 | Visual Instruction Tuning for Qwen2 Base Model | 19 | |
84 | gptparse | Document parser for RAG | 18 | |
85 | evi-CEM | Official implementation of MICCAI2024 paper "Evidential Concept Embedding Models: Towards Reliable Concept Explanations for Skin Disease Diagnosis" | 17 | |
86 | webmarker | Mark web pages for use with vision-language models | 16 | |
87 | LLava-Image-Analyzer | Llava, Ollama and Streamlit | Create POWERFUL Image Analyzer Chatbot for FREE - Windows & Mac | 16 | |
88 | 3d-conditioning | Enhance and modify high-quality compositions using real-time rendering and generative AI output without affecting a hero product asset. | 16 | |
89 | awesome-multilingual-large-language-models | A comprehensive collection of multilingual datasets and large language models, meticulously curated for evaluating and enhancing the performance of large language models across diverse languages and tasks. | 15 | |
90 | ORacle | Official code of the paper ORacle: Large Vision-Language Models for Knowledge-Guided Holistic OR Domain Modeling accepted at MICCAI 2024. | 15 | |
91 | inferenceable | Scalable AI Inference Server for CPU and GPU with Node.js | Utilizes llama.cpp and parts of llamafile C/C++ core under the hood. | 14 | |
92 | HiRED | [AAAI 2025] HiRED strategically drops visual tokens in the image encoding stage to improve inference efficiency for High-Resolution Vision-Language Models (e.g., LLaVA-Next) under a fixed token budget. | 14 | |
93 | Florence-2-Vision-Language-Model | Florence-2 is a novel vision foundation model with a unified, prompt-based representation for a variety of computer vision and vision-language tasks. | 13 | |
94 | VisInContext | Official implementation of Leveraging Visual Tokens for Extended Text Contexts in Multi-Modal Learning | 13 | |
95 | worldcuisines | WorldCuisines is an extensive multilingual and multicultural benchmark that spans 30 languages, covering a wide array of global cuisines. | 13 | |
96 | TIL-2024 | Brainhack TIL 2024: Team 12000SGDPLUSHIE | 12 | |
97 | okra | Okra, your all in one personal AI assistant | 12 | |
98 | WCA | [ICML 2024] Visual-Text Cross Alignment: Refining the Similarity Score in Vision-Language Models | 11 | |
99 | RSTeller | Vision-Language Dataset for Remote Sensing | 11 | |
100 | GVA-Survey | Generalist Virtual Agents: A Survey on Autonomous Agents Across Digital Platforms | 10 |