{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"YOLO-World","owner":"AILab-CVC","isFork":false,"description":"[CVPR 2024] Real-Time Open-Vocabulary Object Detection","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":160,"starsCount":3557,"forksCount":342,"license":"GNU General Public License v3.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,19,6,26,11,14,3,4,16,4,5,10,9,6,2,2,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-20T03:58:08.704Z"}},{"type":"Public","name":"VideoCrafter","owner":"AILab-CVC","isFork":false,"description":"VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models","topicNames":["video-generation","image-to-video","text-to-video"],"topicsNotShown":0,"allTopics":["video-generation","image-to-video","text-to-video"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":47,"starsCount":4135,"forksCount":305,"license":"Other","participation":[0,0,1,0,0,0,0,0,0,0,0,0,2,1,2,0,0,0,0,0,3,2,0,5,5,0,0,1,1,0,0,0,0,0,2,2,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-17T06:20:43.710Z"}},{"type":"Public","name":"SEED-X","owner":"AILab-CVC","isFork":false,"description":"Multimodal Models in Real World","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":3,"starsCount":242,"forksCount":9,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-14T16:17:27.500Z"}},{"type":"Public","name":"SEED-Bench","owner":"AILab-CVC","isFork":false,"description":"(CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":18,"starsCount":257,"forksCount":8,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,8,12,1,9,0,0,2,0,1,0,0,0,0,0,0,0,0,0,27,4,5,5,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,0,0,4,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-14T04:33:42.288Z"}},{"type":"Public","name":"SEED","owner":"AILab-CVC","isFork":false,"description":"Official implementation of SEED-LLaMA (ICLR 2024).","topicNames":["multimodal","vision-language","foundation-model"],"topicsNotShown":0,"allTopics":["multimodal","vision-language","foundation-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":23,"starsCount":485,"forksCount":26,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,3,0,1,0,0,0,1,0,0,1,12,1,22,5,4,2,0,1,1,0,0,0,0,0,0,0,0,0,0,0,4,3,12,3,0,0,0,1,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-11T12:46:51.189Z"}},{"type":"Public","name":"HiFi-123","owner":"AILab-CVC","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":42,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-04T10:55:51.515Z"}},{"type":"Public","name":"AILab-CVC.github.io","owner":"AILab-CVC","isFork":false,"description":"Homepage of Tencent AI Lab CVC.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,2,1,5,4,0,0,0,0,0,6,0,0,0,3,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-28T03:17:47.112Z"}},{"type":"Public","name":"TaleCrafter","owner":"AILab-CVC","isFork":false,"description":"[SIGGRAPH Asia 2023] An interactive story visualization tool that support multiple characters","topicNames":["storytelling","storycreation","siggraph-asia-2023","siggprah-asia"],"topicsNotShown":0,"allTopics":["storytelling","storycreation","siggraph-asia-2023","siggprah-asia"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":5,"starsCount":240,"forksCount":13,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-22T11:53:35.885Z"}},{"type":"Public","name":"M2PT","owner":"AILab-CVC","isFork":false,"description":"[CVPR'24] Multimodal Pathway: Improve Transformers with Irrelevant Data from Other Modalities","topicNames":["deep-learning","transformers","artificial-intelligence","multimodal"],"topicsNotShown":0,"allTopics":["deep-learning","transformers","artificial-intelligence","multimodal"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":69,"forksCount":4,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,2,0,0,0,0,0,0,0,4,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-13T06:28:04.321Z"}},{"type":"Public","name":"Make-Your-Video","owner":"AILab-CVC","isFork":false,"description":"[IEEE TVCG 2024] Customized Video Generation Using Textual and Structural Guidance","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":165,"forksCount":7,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-24T07:56:41.852Z"}},{"type":"Public","name":"Animate-A-Story","owner":"AILab-CVC","isFork":false,"description":"Retrieval-Augmented Video Generation for Telling a Story ","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":2,"starsCount":238,"forksCount":17,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-05T15:07:35.183Z"}},{"type":"Public","name":"FreeNoise","owner":"AILab-CVC","isFork":false,"description":"[ICLR 2024] Code for FreeNoise based on VideoCrafter","topicNames":["generative-model","diffusion","aigc","video-diffusion-model"],"topicsNotShown":0,"allTopics":["generative-model","diffusion","aigc","video-diffusion-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":6,"starsCount":328,"forksCount":24,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-31T08:37:37.334Z"}},{"type":"Public","name":"UniRepLKNet","owner":"AILab-CVC","isFork":false,"description":"[CVPR'24] UniRepLKNet: A Universal Perception Large-Kernel ConvNet for Audio, Video, Point Cloud, Time-Series and Image Recognition","topicNames":["deep-learning","architecture","artificial-intelligence","convolutional-neural-networks","multimodal-learning"],"topicsNotShown":0,"allTopics":["deep-learning","architecture","artificial-intelligence","convolutional-neural-networks","multimodal-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":830,"forksCount":51,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-02T18:17:46.777Z"}},{"type":"Public","name":"GPT4Tools","owner":"AILab-CVC","isFork":false,"description":"GPT4Tools is an intelligent system that can automatically decide, control, and utilize different visual foundation models, allowing the user to interact with images during a conversation.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":14,"starsCount":728,"forksCount":54,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-19T01:55:15.585Z"}},{"type":"Public","name":"VL-GPT","owner":"AILab-CVC","isFork":false,"description":"VL-GPT: A Generative Pre-trained Transformer for Vision and Language Understanding and Generation","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":2,"starsCount":83,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-15T15:01:21.171Z"}},{"type":"Public","name":"GroupMixFormer","owner":"AILab-CVC","isFork":false,"description":"GroupMixAttention and GroupMixFormer","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":107,"forksCount":11,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-13T05:07:44.494Z"}}],"repositoryCount":16,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}