{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"ViP-LLaVA","owner":"WisconsinAIVision","isFork":false,"description":"[CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Prompts","topicNames":["chatbot","llama","multi-modal","clip","vision-language","gpt-4","foundation-models","visual-prompting","llava","llama2"],"topicsNotShown":2,"allTopics":["chatbot","llama","multi-modal","clip","vision-language","gpt-4","foundation-models","visual-prompting","llava","llama2","cvpr2024","gpt-4-vision"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":7,"starsCount":169,"forksCount":9,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-01T18:02:37.307Z"}},{"type":"Public","name":"LLaVA","owner":"WisconsinAIVision","isFork":true,"description":"[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1816,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-18T18:44:01.254Z"}},{"type":"Public","name":"Segment-Everything-Everywhere-All-At-Once","owner":"WisconsinAIVision","isFork":true,"description":"[NeurIPS 2023] Official implementation of the paper \"Segment Everything Everywhere All at Once\"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":329,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-09T06:57:15.744Z"}},{"type":"Public","name":"edit-one-for-all","owner":"WisconsinAIVision","isFork":false,"description":"✏️ Edit One for All: Interactive Batch Image Editing (CVPR 2024)","topicNames":["image-editing","image-manipulation","gans","cvpr","stylegan","stylegan2","batch-image-editor","cvpr2024","batch-image-ed"],"topicsNotShown":0,"allTopics":["image-editing","image-manipulation","gans","cvpr","stylegan","stylegan2","batch-image-editor","cvpr2024","batch-image-ed"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":38,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-11T16:16:13.957Z"}},{"type":"Public","name":"GLIGEN","owner":"WisconsinAIVision","isFork":true,"description":"Open-Set Grounded Text-to-Image Generation","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":140,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-06T06:15:54.697Z"}},{"type":"Public","name":"visii","owner":"WisconsinAIVision","isFork":false,"description":"👀 Visual Instruction Inversion: Image Editing via Visual Prompting (NeurIPS 2023)","topicNames":["image-editing","image-manipulation","neurips","diffusion-models","visual-prompting","neurips-2023"],"topicsNotShown":0,"allTopics":["image-editing","image-manipulation","neurips","diffusion-models","visual-prompting","neurips-2023"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":72,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-19T17:56:19.065Z"}},{"type":"Public","name":"yolact","owner":"WisconsinAIVision","isFork":true,"description":"A simple, fully convolutional model for real-time instance segmentation.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1311,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-06T07:33:12.873Z"}},{"type":"Public","name":"RISE","owner":"WisconsinAIVision","isFork":false,"description":"Domain Generalization through Distilling CLIP with Language Guidance","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":23,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-18T03:20:30.562Z"}},{"type":"Public","name":"UniversalFakeDetect","owner":"WisconsinAIVision","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":10,"starsCount":130,"forksCount":17,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-22T16:22:29.141Z"}},{"type":"Public","name":"few-shot-gan-adaptation","owner":"WisconsinAIVision","isFork":false,"description":"[CVPR '21] Official repository for Few-shot Image Generation via Cross-domain Correspondence ","topicNames":["computer-vision","gans","few-shot-learning"],"topicsNotShown":0,"allTopics":["computer-vision","gans","few-shot-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":22,"starsCount":286,"forksCount":48,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-28T03:20:43.431Z"}},{"type":"Public","name":"INPL","owner":"WisconsinAIVision","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-11T01:48:43.516Z"}},{"type":"Public","name":"W2D","owner":"WisconsinAIVision","isFork":false,"description":"The Two Dimensions of Worst-case Training and the Integrated Effect for Out-of-domain Generalization","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":1,"starsCount":8,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-01-07T05:11:53.873Z"}},{"type":"Public","name":"yolact_edge","owner":"WisconsinAIVision","isFork":false,"description":"The first competitive instance segmentation approach that runs on small edge devices at real-time speeds.","topicNames":["real-time","realtime","pytorch","instance-segmentation","edge-devices","yolactedge"],"topicsNotShown":0,"allTopics":["real-time","realtime","pytorch","instance-segmentation","edge-devices","yolactedge"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":74,"starsCount":1260,"forksCount":275,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-07T02:53:17.238Z"}},{"type":"Public","name":"MaskPoint","owner":"WisconsinAIVision","isFork":false,"description":"[ECCV 2022] Masked Discrimination for Self-Supervised Learning on Point Clouds","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":88,"forksCount":6,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-08-28T01:39:41.888Z"}},{"type":"Public","name":"3detr_MaskPoint","owner":"WisconsinAIVision","isFork":true,"description":"Code & Models for 3DETR - an End-to-end transformer model for 3D object detection","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":76,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-07-31T06:39:02.995Z"}},{"type":"Public","name":"elastic-infogan","owner":"WisconsinAIVision","isFork":false,"description":"Official PyTorch implementation of Elastic-InfoGAN [NeurIPS 2020]","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":13,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-04-03T20:19:30.167Z"}},{"type":"Public","name":"MixNMatch","owner":"WisconsinAIVision","isFork":false,"description":"Pytorch implementation of MixNMatch","topicNames":["deep-learning","image-manipulation","image-generation","gans","fine-grained","disentangled-representations","pytorch"],"topicsNotShown":0,"allTopics":["deep-learning","image-manipulation","image-generation","gans","fine-grained","disentangled-representations","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":12,"starsCount":978,"forksCount":190,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-07-07T23:01:18.345Z"}}],"repositoryCount":17,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}