Public code of the ML course
-
Updated
Jun 7, 2017 - Python
Public code of the ML course
[ICML 2022] "Data-Efficient Double-Win Lottery Tickets from Robust Pre-training" by Tianlong Chen, Zhenyu Zhang, Sijia Liu, Yang Zhang, Shiyu Chang, Zhangyang Wang
3D Loss Landscapes of SoftNet (Sparse Subnetwork)
Minimal Reproducibility Study of (https://arxiv.org/abs/1911.05248). Experiments with Compression of Deep Neural Networks
Project examing sparse deep learning architectures for ligand classification.
Official implementation of the paper "HyperSparse Neural Networks: Shifting Exploration to Exploitation through Adaptive Regularization"
A pure implementation for sparse denoising autoencoder with adaptive evolutionary training using Scipy. The sparse implementation makes the algorithm scalable to high dimensional data and trainable on CPUs.
Code to reproduce the experiments of the ICLR24-paper: "Sparse Model Soups: A Recipe for Improved Pruning via Model Averaging"
analysing Model Pruning and Unit Pruning on a large dense MNIST network
[ICCV2023 Official PyTorch code] for Iterative Soft Shrinkage Learning for Efficient Image Super-Resolution
Code for CPAL-2024 paper "Continual Learning with Dynamic Sparse Training: Exploring Algorithms for Effective Model Updates"
[NeurIPS 2022] "Sparse Winning Tickets are Data-Efficient Image Recognizers" by Mukund Varma T, Xuxi Chen, Zhenyu Zhang, Tianlong Chen, Subhashini Venugopalan, Zhangyang Wang
Generalized Orthogonal Least-Squares in CUDA
Repository containing Parameter Inference and Posterior Computation codes of CMAP
WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.
Molecular-property prediction with sparsity
[AAMAS 2023] Code for the paper "Automatic Noise Filtering with Dynamic Sparse Training in Deep Reinforcement Learning"
A simple C++14 and CUDA-based header-only library with tools for sparse-machine learning.
Learning to Rearrange Voxels in Binary Segmentation Masks for Smooth Manifold Triangulation
Add a description, image, and links to the sparsity topic page so that developers can more easily learn about it.
To associate your repository with the sparsity topic, visit your repo's landing page and select "manage topics."