Here are
439 public repositories
matching this topic...
ncnn is a high-performance neural network inference framework optimized for the mobile platform
MediaPipe is the simplest way for researchers and developers to build world-class ML solutions and applications for mobile, edge, cloud and the web.
YOLOv3 in PyTorch > ONNX > CoreML > iOS
-
Updated
Jul 20, 2020
-
Jupyter Notebook
💎1MB lightweight face detection model (1MB轻量级人脸检测模型)
-
Updated
Jun 20, 2020
-
Python
Hello AI World guide to deploying deep-learning inference networks and deep vision primitives with TensorRT and NVIDIA Jetson.
Runtime type system for IO decoding/encoding
-
Updated
Jul 21, 2020
-
TypeScript
Grakn Core: The Knowledge Graph
-
Updated
Jul 23, 2020
-
Java
TensorFlow template application for deep learning
-
Updated
Jan 3, 2019
-
Python
OpenVINO™ Toolkit repository
An easy to use PyTorch to TensorRT converter
-
Updated
Jul 8, 2020
-
Python
Acceleration package for neural networks on multi-core CPUs
The Triton Inference Server provides a cloud inferencing solution optimized for NVIDIA GPUs.
DELTA is a deep learning based natural language and speech processing platform.
-
Updated
Jul 20, 2020
-
Python
A curated list of research in machine learning system. I also summarize some papers if I think they are really interesting.
A uniform interface to run deep learning models from multiple frameworks
Multi Model Server is a tool for serving neural net models for inference
-
Updated
Jul 22, 2020
-
Java
Pytorch-Named-Entity-Recognition-with-BERT
-
Updated
Jan 24, 2020
-
Python
The challenge projects for Inferencing machine learning models on iOS
-
Updated
Nov 26, 2019
-
Python
TensorFlow models accelerated with NVIDIA TensorRT
-
Updated
Nov 22, 2019
-
Python
Embedded and mobile deep learning research resources
High-efficiency floating-point neural network inference operators for mobile, server, and Web
Shape inference for PyTorch (like Keras) & new layers
-
Updated
May 30, 2020
-
Python
a fast and user-friendly runtime for transformer inference on CPU and GPU
A curated list of awesome TensorFlow Lite models, samples, tutorials, tools and learning resources.
A REST API for Caffe using Docker and Go
TensorFlow examples in C, C++, Go and Python without bazel but with cmake and FindTensorFlow.cmake
-
Updated
Aug 18, 2019
-
CMake
Lua Language Server coded by Lua
A scalable inference server for models optimized with OpenVINO™
-
Updated
Jul 22, 2020
-
Python
Fast implementation of BERT inference directly on NVIDIA (CUDA, CUBLAS) and Intel MKL
Improve this page
Add a description, image, and links to the
inference
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
inference
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.
You signed in with another tab or window. Reload to refresh your session.
You signed out in another tab or window. Reload to refresh your session.