This is the Linux app named AIMET whose latest release can be downloaded as 1.28.0.zip. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named AIMET with OnWorks for free.
Follow these instructions in order to run this app:
- 1. Downloaded this application in your PC.
- 2. Enter in our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 3. Upload this application in such filemanager.
- 4. Start the OnWorks Linux online or Windows online emulator or MACOS online emulator from this website.
- 5. From the OnWorks Linux OS you have just started, goto our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 6. Download the application, install it and run it.
SCREENSHOTS
Ad
AIMET
DESCRIPTION
Qualcomm Innovation Center (QuIC) is at the forefront of enabling low-power inference at the edge through its pioneering model-efficiency research. QuIC has a mission to help migrate the ecosystem toward fixed-point inference. With this goal, QuIC presents the AI Model Efficiency Toolkit (AIMET) - a library that provides advanced quantization and compression techniques for trained neural network models. AIMET enables neural networks to run more efficiently on fixed-point AI hardware accelerators. Quantized inference is significantly faster than floating point inference. For example, models that we’ve run on the Qualcomm® Hexagon™ DSP rather than on the Qualcomm® Kryo™ CPU have resulted in a 5x to 15x speedup. Plus, an 8-bit model also has a 4x smaller memory footprint relative to a 32-bit model. However, often when quantizing a machine learning model (e.g., from 32-bit floating point to an 8-bit fixed point value), the model accuracy is sacrificed.
Features
- Equalize weight tensors to reduce amplitude variation across channels
- Tensor-decomposition technique to split a large layer into two smaller ones
- Corrects shift in layer outputs introduced due to quantization
- Removes redundant input channels from a layer and reconstructs layer weights
- Use quantization sim to train the model further to improve accuracy
- Automatically selects how much to compress each layer in the model
Programming Language
Python
Categories
This is an application that can also be fetched from https://sourceforge.net/projects/aimet.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.