Skip to content
#

tensorrt

Here are 365 public repositories matching this topic...

TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is distinguished by several outstanding features, including its cross-platform capability, high performance, model compression and code pruning. Based on ncnn and Rapidnet, TNN further strengthens the support and performance optimization for mobile devices, and also draws on the advantages of good extensibility and high performance from existed open source efforts. TNN has been deployed in multiple Apps from Tencent, such as Mobile QQ, Weishi, Pitu, etc. Contributions are welcome to work in collaborative with us and make TNN a better framework.

  • Updated Jun 6, 2022
  • C++

micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape

  • Updated Oct 6, 2021
  • Python
morgoth95
morgoth95 commented Apr 24, 2022

The latest release of openvino has changed the inference engine API, providing a new API that takes advantage of the full potential of Intel's latest version of IR (IR 11). More information can be found at this link. We should adapt nebullvm's OpenVinoInferenceLearner to their latest API.

Usef

enhancement good first issue

Improve this page

Add a description, image, and links to the tensorrt topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the tensorrt topic, visit your repo's landing page and select "manage topics."

Learn more