ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
machine-learning
deep-learning
pytorch
tensorflow
neural-networks
scikit-learn
ai-framework
hardware-acceleration
onnx
Обновлено 2024-09-17 22:00:24 +03:00
Simple and Distributed Machine Learning
azure
microsoft
machine-learning
deep-learning
ai
data-science
opencv
ml
cognitive-services
spark
http
big-data
lightgbm
databricks
onnx
apache-spark
pyspark
scala
model-deployment
synapse
Обновлено 2024-09-16 09:47:20 +03:00
The pre- and post- processing library for ONNX Runtime
Обновлено 2024-09-16 08:34:07 +03:00
The pre- and post- processing library for ONNX Runtime
Обновлено 2024-09-16 08:34:07 +03:00
Examples for using ONNX Runtime for machine learning inferencing.
Обновлено 2024-09-04 02:29:49 +03:00
Examples for using ONNX Runtime for model training.
Обновлено 2024-08-09 21:26:43 +03:00
A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.
machine-learning
python
deep-learning
pytorch
neural-architecture-search
deep-neural-networks
edge-computing
latency
inference
edge-ai
efficient-model
onnx-models
tensorflow-models
Обновлено 2024-07-31 00:16:53 +03:00
Common utilities for ONNX converters
Обновлено 2024-06-14 03:56:41 +03:00
Samples and Tools for Windows ML.
machine-learning
deep-learning
windows
pytorch
ai
ml
tensorflow
scikit-learn
neural-network
onnx
keras
coreml
caffe2
winmltools
Обновлено 2024-05-28 23:28:15 +03:00
Обновлено 2024-04-15 23:26:33 +03:00
ONNX Runtime Web benchmark tool
Обновлено 2023-08-16 03:11:17 +03:00
Обновлено 2023-07-12 10:36:13 +03:00
demos to show the capabilities of ONNX Runtime Web
Обновлено 2023-07-08 22:25:38 +03:00
PyTorch ObjectDetection Modules and ONNX ops
Обновлено 2023-06-12 21:23:12 +03:00
Open Enclave port of the ONNX runtime for confidential inferencing on Azure Confidential Computing
Обновлено 2022-12-12 19:28:21 +03:00
Open deep learning compiler stack for cpu, gpu and specialized accelerators
Обновлено 2022-11-28 22:09:42 +03:00
MMdnn is a set of tools to help users inter-operate among different deep learning frameworks. E.g. model conversion and visualization. Convert models between Caffe, Keras, MXNet, Tensorflow, CNTK, PyTorch Onnx and CoreML.
Обновлено 2022-09-23 02:59:07 +03:00
An Open Enclave port of the ONNX inference server with data encryption and attestation capabilities to enable confidential inference on Azure Confidential Computing.
Обновлено 2022-08-29 19:33:49 +03:00
demos to show the capabilities of ONNX.js
Обновлено 2022-01-29 00:42:51 +03:00
ONNX.js: run ONNX models using JavaScript
Обновлено 2021-12-14 03:51:01 +03:00
Обновлено 2021-11-19 02:56:39 +03:00