{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"torchquantum","owner":"mit-han-lab","isFork":false,"description":"A PyTorch-based framework for Quantum Classical Simulation, Quantum Machine Learning, Quantum Neural Networks, Parameterized Quantum Circuits with support for easy deployments on real quantum computers.","allTopics":["machine-learning","system","deep-learning","neural-network","quantum","pytorch","quantum-computing","quantum-machine-learning","quantum-simulation","ml-for-systems","pytorch-quantum","quantum-neural-network","parameterized-quantum-circuit"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":12,"issueCount":61,"starsCount":1219,"forksCount":180,"license":"MIT License","participation":[48,0,1,1,4,2,6,10,0,5,12,36,0,5,1,0,1,11,7,1,9,3,4,21,9,2,0,12,3,0,2,13,7,0,3,4,20,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-11T19:29:16.915Z"}},{"type":"Public","name":"litepose","owner":"mit-han-lab","isFork":false,"description":"[CVPR'22] Lite Pose: Efficient Architecture Design for 2D Human Pose Estimation","allTopics":["pose-estimation","efficient-models","litepose"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":18,"starsCount":297,"forksCount":35,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:04:06.770Z"}},{"type":"Public","name":"gan-compression","owner":"mit-han-lab","isFork":false,"description":"[CVPR 2020] GAN Compression: Efficient Architectures for Interactive Conditional GANs","allTopics":["compression","pytorch","gans","pix2pix","cyclegan","image-to-image-translation","conditional-gans","gaugan"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":3,"starsCount":1097,"forksCount":147,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T21:59:17.248Z"}},{"type":"Public","name":"TinyChatEngine","owner":"mit-han-lab","isFork":false,"description":"TinyChatEngine: On-Device LLM Inference Library","allTopics":["c","arm","deep-learning","cpp","x86-64","quantization","edge-computing","cuda-programming","on-device-ai","large-language-models"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":2,"issueCount":26,"starsCount":584,"forksCount":56,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,6,4,4,5,1,2,3,1,5,0,2,0,0,2,0,0,0,0,0,1,0,3,0,3,0,1,3,3,0,0,0,0,0,0,2,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T05:28:31.231Z"}},{"type":"Public","name":"llm-awq","owner":"mit-han-lab","isFork":false,"description":"[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":111,"starsCount":2002,"forksCount":140,"license":"MIT License","participation":[0,6,5,7,1,17,3,6,2,3,0,5,3,1,0,4,0,0,1,10,5,2,0,0,5,0,0,0,0,0,0,0,0,0,0,0,9,3,2,0,0,1,0,0,0,1,3,1,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T04:29:53.463Z"}},{"type":"Public","name":"efficientvit","owner":"mit-han-lab","isFork":false,"description":"EfficientViT is a new family of vision models for efficient high-resolution vision.","allTopics":["imagenet","segmentation","high-resolution","vision-transformer","efficientvit","segment-anything"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":74,"starsCount":1528,"forksCount":136,"license":"Apache License 2.0","participation":[0,0,2,0,0,1,0,0,0,0,1,0,5,8,9,3,1,0,1,2,0,0,1,1,0,0,0,0,1,0,0,1,0,0,5,4,1,0,5,5,4,2,3,0,0,3,2,4,2,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T15:58:03.753Z"}},{"type":"Public","name":"distrifuser","owner":"mit-han-lab","isFork":false,"description":"[CVPR 2024 Highlight] DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models","allTopics":["acceleration","parallelism","generative-model","diffusion-models","generative-ai"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":460,"forksCount":12,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,10,1,0,0,0,2,0,0,4,3,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T05:20:37.304Z"}},{"type":"Public","name":"torchsparse","owner":"mit-han-lab","isFork":false,"description":"[MICRO'23, MLSys'22] TorchSparse: Efficient Training and Inference Framework for Sparse Convolution on GPUs.","allTopics":["acceleration","pytorch"],"primaryLanguage":{"name":"Cuda","color":"#3A4E3A"},"pullRequestCount":0,"issueCount":23,"starsCount":1136,"forksCount":128,"license":"MIT License","participation":[4,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T02:03:43.161Z"}},{"type":"Public","name":"patch_conv","owner":"mit-han-lab","isFork":false,"description":"Patch convolution to avoid large GPU memory usage of Conv2D","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":1,"starsCount":63,"forksCount":4,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,11,0,0,0,2,0,0,0,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-26T05:14:52.126Z"}},{"type":"Public","name":"qserve","owner":"mit-han-lab","isFork":false,"description":"QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":14,"starsCount":282,"forksCount":6,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-14T16:49:05.246Z"}},{"type":"Public","name":"lmquant","owner":"mit-han-lab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":63,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-11T17:26:11.936Z"}},{"type":"Public","name":"spatten-llm","owner":"mit-han-lab","isFork":false,"description":"[HPCA'21] SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning","allTopics":["rtl","attention","hardware-acceleration","spinalhdl","llm-inference"],"primaryLanguage":{"name":"Scala","color":"#c22d40"},"pullRequestCount":0,"issueCount":1,"starsCount":52,"forksCount":3,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-03T21:28:22.876Z"}},{"type":"Public","name":"smoothquant","owner":"mit-han-lab","isFork":false,"description":"[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":55,"starsCount":1074,"forksCount":123,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,3,1,0,0,0,9,0,0,0,0,0,0,0,0,0,0,0,1,0,6,0,0,1,1,0,0,0,1,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-28T17:17:11.771Z"}},{"type":"Public","name":"sparsevit","owner":"mit-han-lab","isFork":false,"description":"[CVPR'23] SparseViT: Revisiting Activation Sparsity for Efficient High-Resolution Vision Transformer","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":55,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-24T01:30:59.362Z"}},{"type":"Public","name":"mcunet","owner":"mit-han-lab","isFork":false,"description":"[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning","allTopics":["deep-learning","pytorch","neural-architecture-search","tinyml","microncontroller"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":20,"starsCount":416,"forksCount":77,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-29T18:17:37.080Z"}},{"type":"Public","name":"tinyengine","owner":"mit-han-lab","isFork":false,"description":"[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning; [NeurIPS 2022] MCUNetV3: On-Device Training Under 256KB Memory","allTopics":["c","microcontroller","cpp","pytorch","codegenerator","tinyml","deep-learning","quantization","edge-computing","neural-architecture-search"],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":1,"issueCount":31,"starsCount":754,"forksCount":126,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,0,0,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-29T18:16:51.627Z"}},{"type":"Public","name":"tiny-training","owner":"mit-han-lab","isFork":false,"description":"On-Device Training Under 256KB Memory [NeurIPS'22]","allTopics":["edge-ai","on-device-training","learning-on-the-edge"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":7,"starsCount":409,"forksCount":54,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-29T18:15:19.456Z"}},{"type":"Public","name":"bevfusion","owner":"mit-han-lab","isFork":false,"description":"[ICRA'23] BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation","allTopics":["camera","pytorch","lidar","object-detection","sensor-fusion","semantic-segmentation","3d-perception"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":54,"starsCount":2086,"forksCount":377,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-29T00:41:58.791Z"}},{"type":"Public","name":"streaming-llm","owner":"mit-han-lab","isFork":false,"description":"[ICLR 2024] Efficient Streaming Language Models with Attention Sinks","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":35,"starsCount":6305,"forksCount":356,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,14,10,0,3,3,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-20T05:46:26.421Z"}},{"type":"Public","name":"tinychat-tutorial","owner":"mit-han-lab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":1,"issueCount":3,"starsCount":34,"forksCount":9,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-01T23:32:08.814Z"}},{"type":"Public","name":"once-for-all","owner":"mit-han-lab","isFork":false,"description":"[ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment","allTopics":["acceleration","nas","automl","edge-ai","efficient-model","tinyml"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":53,"starsCount":1846,"forksCount":332,"license":"MIT License","participation":[0,0,0,0,0,6,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-14T04:44:52.937Z"}},{"type":"Public","name":"fastcomposer","owner":"mit-han-lab","isFork":false,"description":"FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":15,"starsCount":613,"forksCount":34,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-09T03:14:15.846Z"}},{"type":"Public","name":"tinyml","owner":"mit-han-lab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":6,"starsCount":716,"forksCount":130,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-29T04:22:13.211Z"}},{"type":"Public","name":"offsite-tuning","owner":"mit-han-lab","isFork":false,"description":"Offsite-Tuning: Transfer Learning without Full Model","allTopics":["deep-learning","transfer-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":5,"starsCount":362,"forksCount":36,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-27T17:39:38.113Z"}},{"type":"Public","name":"amc","owner":"mit-han-lab","isFork":false,"description":"[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices","allTopics":["automl","model-compression","channel-pruning","automl-for-compression","efficient-model","on-device-ai"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":16,"starsCount":420,"forksCount":108,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-22T23:33:25.504Z"}},{"type":"Public","name":"data-efficient-gans","owner":"mit-han-lab","isFork":false,"description":"[NeurIPS 2020] Differentiable Augmentation for Data-Efficient GAN Training","allTopics":["tensorflow","generative-adversarial-network","image-generation","gans","data-efficient","neurips-2020","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":26,"starsCount":1267,"forksCount":175,"license":"BSD 2-Clause \"Simplified\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-12T00:16:36.633Z"}},{"type":"Public","name":"flatformer","owner":"mit-han-lab","isFork":false,"description":"[CVPR'23] FlatFormer: Flattened Window Attention for Efficient Point Cloud Transformer","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":108,"forksCount":12,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-27T04:02:56.896Z"}},{"type":"Public","name":"anycost-gan","owner":"mit-han-lab","isFork":false,"description":"[CVPR 2021] Anycost GANs for Interactive Image Synthesis and Editing","allTopics":["computer-vision","deep-learning","computer-graphics","pytorch","image-editing","generative-adversarial-network","gan","image-manipulation","image-generation","gans","stylegan2"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":772,"forksCount":96,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-03T15:20:42.622Z"}},{"type":"Public","name":"temporal-shift-module","owner":"mit-han-lab","isFork":false,"description":"[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding","allTopics":["acceleration","low-latency","video-understanding","efficient-model","temporal-modeling","tsm","nvidia-jetson-nano"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":93,"starsCount":2030,"forksCount":417,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-03T15:14:16.654Z"}},{"type":"Public","name":"proxylessnas","owner":"mit-han-lab","isFork":false,"description":"[ICLR 2019] ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware","allTopics":["acceleration","automl","specialization","efficient-model","on-device-ai","hardware-aware"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":1413,"forksCount":283,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-26T01:16:43.630Z"}}],"repositoryCount":50,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}