<p align="center"><img width="50%" src="docs/images/ONNX_Runtime_logo_dark.png" /></p>
**ONNX Runtime is a cross-platform inference and training machine-learning accelerator**.
**ONNX Runtime inference** can enable faster customer experiences and lower costs, supporting models from deep learning frameworks such as PyTorch and TensorFlow/Keras as well as classical machine learning libraries such as scikit-learn, LightGBM, XGBoost, etc. ONNX Runtime is compatible with different hardware, drivers, and operating systems, and provides optimal performance by leveraging hardware accelerators where applicable alongside graph optimizations and transforms. [Learn more →](https://www.onnxruntime.ai/docs/#onnx-runtime-for-inferencing)
**ONNX Runtime training** can accelerate the model training time on multi-node NVIDIA GPUs for transformer models with a one-line addition for existing PyTorch training scripts. [Learn more →](https://www.onnxruntime.ai/docs/#onnx-runtime-for-training)
## Get Started & Resources
* **General Information**: [onnxruntime.ai](https://onnxruntime.ai)
* **Usage documentation and tutorials**: [onnxruntime.ai/docs](https://onnxruntime.ai/docs)
* **YouTube video tutorials**: [youtube.com/@ONNXRuntime](https://www.youtube.com/@ONNXRuntime)
* [**Upcoming Release Roadmap**](https://github.com/microsoft/onnxruntime/wiki/Upcoming-Release-Roadmap)
* **Companion sample repositories**:
- ONNX Runtime Inferencing: [microsoft/onnxruntime-inference-examples](https://github.com/microsoft/onnxruntime-inference-examples)
- ONNX Runtime Training: [microsoft/onnxruntime-training-examples](https://github.com/microsoft/onnxruntime-training-examples)
## Builtin Pipeline Status
|System|Inference|Training|
|---|---|---|
|Windows|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Windows%20CPU%20CI%20Pipeline?label=Windows+CPU)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=9)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Windows%20GPU%20CI%20Pipeline?label=Windows+GPU)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=10)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Windows%20GPU%20TensorRT%20CI%20Pipeline?label=Windows+GPU+TensorRT)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=47)||
|Linux|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Linux%20CPU%20CI%20Pipeline?label=Linux+CPU)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=11)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Linux%20CPU%20Minimal%20Build%20E2E%20CI%20Pipeline?label=Linux+CPU+Minimal+Build)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=64)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Linux%20GPU%20CI%20Pipeline?label=Linux+GPU)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=12)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Linux%20GPU%20TensorRT%20CI%20Pipeline?label=Linux+GPU+TensorRT)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=45)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Linux%20OpenVINO%20CI%20Pipeline?label=Linux+OpenVINO)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=55)|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/orttraining-linux-ci-pipeline?label=Linux+CPU+Training)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=86)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/orttraining-linux-gpu-ci-pipeline?label=Linux+GPU+Training)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=84)<br>[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/orttraining/orttraining-ortmodule-distributed?label=Training+Distributed)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=148)|
|Mac|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/MacOS%20CI%20Pipeline?label=MacOS+CPU)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=13)||
|Android|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/Android%20CI%20Pipeline?label=Android)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=53)||
|iOS|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/iOS%20CI%20Pipeline?label=iOS)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=134)||
|Web|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/ONNX%20Runtime%20Web%20CI%20Pipeline?label=Web)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=161)||
|Other|[![Build Status](https://dev.azure.com/onnxruntime/onnxruntime/_apis/build/status/onnxruntime-binary-size-checks-ci-pipeline?repoName=microsoft%2Fonnxruntime&label=Binary+Size+Check)](https://dev.azure.com/onnxruntime/onnxruntime/_build/latest?definitionId=187&repoName=microsoft%2Fonnxruntime)||
## Third-party Pipeline Status
|System|Inference|Training|
|---|---|---|
|Linux|[![Build Status](https://github.com/Ascend/onnxruntime/actions/workflows/build-and-test.yaml/badge.svg)](https://github.com/Ascend/onnxruntime/actions/workflows/build-and-test.yaml)||
## Data/Telemetry
Windows distributions of this project may collect usage data and send it to Microsoft to help improve our products and services. See the [privacy statement](docs/Privacy.md) for more details.
## Contributions and Feedback
We welcome contributions! Please see the [contribution guidelines](CONTRIBUTING.md).
For feature requests or bug reports, please file a [GitHub Issue](https://github.com/Microsoft/onnxruntime/issues).
For general discussion or questions, please use [GitHub Discussions](https://github.com/microsoft/onnxruntime/discussions).
## Code of Conduct
This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/).
For more information see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq/)
or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
## License
This project is licensed under the [MIT License](LICENSE).
没有合适的资源?快使用搜索试试~ 我知道了~
LaMa Image Inpainting 图像修复 OnnxRuntime-GPU版 Demo.rar
共219个文件
dll:49个
_:41个
xml:30个
1.该资源内容由用户上传,如若侵权请联系客服进行举报
2.虚拟产品一经售出概不退款(资源遇到问题,请及时私信上传者)
2.虚拟产品一经售出概不退款(资源遇到问题,请及时私信上传者)
版权申诉
0 下载量 9 浏览量
2024-07-01
12:11:22
上传
评论
收藏 276.02MB RAR 举报
温馨提示
LaMa Image Inpainting 图像修复 OnnxRuntime-GPU版 Demo.rar 博客地址:https://lw112190.blog.csdn.net/article/details/140096954
资源推荐
资源详情
资源评论
收起资源包目录
LaMa Image Inpainting 图像修复 OnnxRuntime-GPU版 Demo.rar (219个子文件)
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
_._ 0B
packages.config 844B
app.config 781B
Form1.Designer.cs 6KB
Form1.cs 3KB
Resources.Designer.cs 3KB
Common.cs 3KB
LaMa.cs 2KB
AssemblyInfo.cs 1KB
Settings.Designer.cs 1KB
Program.cs 536B
LaMa Image Inpainting 图像修复 OnnxRuntime-GPU版 Demo.csproj 9KB
onnxruntime_providers_cuda.dll 392.82MB
OpenCvSharpExtern.dll 57.96MB
OpenCvSharpExtern.dll 39.72MB
opencv_videoio_ffmpeg480_64.dll 25.13MB
opencv_videoio_ffmpeg480.dll 22.3MB
onnxruntime.dll 11.1MB
OpenCvSharp.dll 922KB
OpenCvSharp.dll 919KB
OpenCvSharp.dll 919KB
OpenCvSharp.dll 919KB
onnxruntime_providers_tensorrt.dll 700KB
Microsoft.ML.OnnxRuntime.dll 205KB
Microsoft.ML.OnnxRuntime.dll 204KB
Microsoft.ML.OnnxRuntime.dll 204KB
Microsoft.ML.OnnxRuntime.dll 204KB
Microsoft.ML.OnnxRuntime.dll 202KB
Microsoft.ML.OnnxRuntime.dll 202KB
Microsoft.ML.OnnxRuntime.dll 202KB
Microsoft.ML.OnnxRuntime.dll 201KB
System.Numerics.Vectors.dll 160KB
System.Numerics.Vectors.dll 157KB
System.Numerics.Vectors.dll 157KB
System.Memory.dll 139KB
System.Memory.dll 139KB
System.Memory.dll 135KB
System.Numerics.Vectors.dll 113KB
zlibwapi.dll 87KB
System.ValueTuple.dll 78KB
System.ValueTuple.dll 77KB
System.ValueTuple.dll 77KB
System.ValueTuple.dll 41KB
System.ValueTuple.dll 40KB
System.Numerics.Vectors.dll 37KB
System.Numerics.Vectors.dll 37KB
System.Numerics.Vectors.dll 37KB
System.Numerics.Vectors.dll 29KB
System.ValueTuple.dll 25KB
onnxruntime_providers_shared.dll 21KB
System.ValueTuple.dll 21KB
System.Buffers.dll 21KB
System.Buffers.dll 20KB
System.Buffers.dll 20KB
System.Runtime.CompilerServices.Unsafe.dll 18KB
System.Runtime.CompilerServices.Unsafe.dll 18KB
System.Runtime.CompilerServices.Unsafe.dll 18KB
System.Runtime.CompilerServices.Unsafe.dll 18KB
System.Buffers.dll 14KB
System.Buffers.dll 14KB
共 219 条
- 1
- 2
- 3
资源评论
天天代码码天天
- 粉丝: 1w+
- 资源: 606
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
最新资源
- GAPSO-LSTM,即遗传粒子群优化算法优化LSTM的超参数做数据回归预测,多输入单输出,预测精度高于PSO-LSTM,算法原
- 永磁同步电机的参数辨识源码,完整的CCS工程,已经在工程项目上验证通过,辨识精度非常高 1、参数辨识源码在src-foc文件夹
- 增程式电动汽车基于工况的自适应ECMS能量管理策略(matlab的m程序)
- Fluent电弧,激光,熔滴一体模拟 UDF包括高斯旋转体热源、双椭球热源(未使用)、VOF梯度计算、反冲压力、磁场力、表面张
- C#全自动多线程上位机源码编程 0,纯源代码 1,替代传统plc搭载的触摸屏 2,工控屏幕一体机直接和plc通信 3,功能
- 基于三有源桥的模型预测控制仿真,可以独立控制输出侧两个端口的电压或者电流,动态响应快,也可以扩展至四有源桥电路
- VIENNA维也纳拓扑,三相整流simulink仿真:采用电压电流双闭环控制,电压外环采用PI控制,电流内环采用bang ban
- 永磁同步电机改进超螺旋滑模观测器无位置传感器控制 采用一种改进的超螺旋滑模观测器永磁同步电机无位置传感器控制,该观测器在传统ST
- comsol仿真模拟气液两相化学吸收CO2(氢氧化钠溶液NaOH和MEA溶液吸收CO2) 此案例为文献复现
- 内有cpar文件和simulink文件,并有演示操作视频,carsim+simulink联合仿真实实现道超车, 包含道决策,路径
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功