Collection of recent methods on DNN compression and acceleration
The developer of this repository has not created any items for sale yet. Need a bug fixed? Help with integration? A different license? Create a request here:
A collection of recent methods on DNN compression and acceleration. There are mainly 5 kinds of methods for efficient DNNs: - neural architecture re-designing or searching - maintain accuracy, less cost (e.g., #Params, #FLOPs, etc.): MobileNet, ShuffleNet etc. - maintain cost, more accuracy: Inception, ResNeXt, Xception etc. - pruning (including structured and unstructured) - quantization - matrix decomposition - knowledge distillation
About abbreviation: In the list below,ofor oral,wfor workshop,sfor spotlight,bfor best paper.