You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
ONNX Model Zoo Demo | Tutorial-10 | Open Neural Network Exchange | ONNX
ONNX Model Zoo Demo | Tutorial-10 | Open Neural Network Exchange | ONNX
The video tutorial showcases how to use ONNX Model Zoo to perform inferencing on an ONNX model using the ONNX runtime. The presenter guides viewers through the process of creating a virtual environment, installing necessary packages, downloading the MNIST handwritten model from the ONNX Model Zoo, and writing a Python script for inference. The demo shows that the prediction time is fast and encourages users to download models directly from the ONNX Model Zoo. The video teases the next tutorial, which will cover converting a Python model to TensorFlow.
PyTorch to Tensorflow Demo | Tutorial-11 | Open Neural Network Exchange | ONNX
PyTorch to Tensorflow Demo | Tutorial-11 | Open Neural Network Exchange | ONNX
The video demonstrates how to use ONNX to convert a PyTorch model to TensorFlow format. The process involves training the model in PyTorch, saving it in .pth format, and then converting it to ONNX format before finally converting it to TensorFlow format. The conversion process is shown in detail through the use of a handwritten digit classification model using the MNIST dataset, and the resulting TensorFlow model is tested with sample images. The video also briefly touches upon converting a model from Caffe2 to ONNX and suggests that users explore ONNX further.functions. The notebook code is provided in the resources section for users to follow along.
Netron is a tool for viewing neural network, deep learning, and machine learning models
Quick look into Netron
Quick look into Netron
In the video, the presenter provides an overview of Netron, a tool for viewing and analyzing machine learning models. Netron supports various formats and can be installed on multiple platforms. The presenter demonstrates how to start Netron and navigate through several example models, highlighting the tool's capabilities and limitations. While Netron is useful for exploring simpler network architectures, the presenter suggests that it could benefit from additional features for visualizing more complex models. Overall, the presenter recommends Netron as a helpful tool for examining and understanding machine learning models.
Netron - Network Visualization Tool | Machine Learning | Data Magic
Netron - Network Visualization Tool | Machine Learning | Data Magic
Netron is a Python library that helps users to visually explore and examine the structure and parameters of deep learning models. It is an open-source library that provides sample models for analysis and has a simple installation process. With just two lines of code, users can install Netron and use it to visualize the neural network structure, activation functions, pooling layers, convolutional layers, and all attributes passed at each layer of a given machine learning model. Netron provides an easy-to-use interface that allows users to export visualizations as PNG files and explore different features and options.
PyTorch, TensorFlow, Keras, ONNX, TensorRT, OpenVINO, AI Model File Conversion
[Educational Video] PyTorch, TensorFlow, Keras, ONNX, TensorRT, OpenVINO, AI Model File Conversion
The speaker in the video discusses the advantages and trade-offs of different AI frameworks, such as PyTorch, TensorFlow, Keras, ONNX, TensorRT, and OpenVINO, and recommends PyTorch as the preferred framework for training and data conversion. The speaker explains the conversion process, including converting PyTorch models to ONNX and then to TensorRT or OpenVINO, and cautions against using TensorFlow PB file and Cafe. The speaker also discusses the importance of setting the floating point format properly and recommends using FP 32 for most models. The video provides examples of model conversion and encourages viewers to visit the official website for more educational videos.
How we use ONNX in Zetane to complete machine learning projects faster with less trial-and-error
How we use ONNX in Zetane to complete machine learning projects faster with less trial-and-error
Patrick Saitama, co-founder and CTO of Zetane Systems, discusses the value of using ONNX in his company's new product to address issues related to the black box problem of AI. Zetane's engine allows for the exploration and inspection of the ONNX models, providing insights into the model's interaction with data and leading to more decisive strategies for improving its quality. The example given shows how Zetane's engine helped debug an autonomous train model by inspecting the radio layer and adding more images of tunnels labeled as non-obstacles. Zetane also includes tools for dynamically inspecting internal tensors and taking snapshots of the model for later investigation. Additionally, Zetane's new engine allows for larger models such as YOLOv3 to be installed.
What's New in ONNX Runtime
What's New in ONNX Runtime
This talk will share highlights of the ONNX Runtime 1.10-1.12 releases, including details on notable performance improvements, features, and platforms including mobile and web. Ryan Hill has been with the AI Frameworks team for the past 4 years, where he has mostly worked on operator kernels, C APIs, and dynamically loading execution providers. Prior to this he worked on the Office PowerPoint team, where his most widely seen work is many of the slideshow slide transitions. For fun he likes trying to use the latest C++ features and hitting internal compiler errors.In the video, software engineer Ryan Hill discusses the various features and updates of ONNX Runtime, a widely used cross-platform runtime that can target multiple CPU architectures. He highlights the latest features added to ONNX Runtime, such as the ability to call op kernels directly and performance improvements like transpose optimizer and small size optimization. Hill also talks about ONNX Runtime's execution providers, which enable optimal performance on various hardware, and the release of mobile packages that support NHWC conversion at runtime. The video also covers layout-sensitive operator support, Xamarin support for cross-platform apps, ONNX Runtime web, and the ONNX Runtime extensions library that focuses on model pre-post-processing work, including text conversions and mathematical operations, and currently focuses on NLP, vision, and text domains.
v1.12.0 ONNX Runtime - Release Review
v1.12.0 ONNX Runtime - Release Review
The v1.12.0 release of the ONNX Runtime (ORT) focuses on inferencing but also includes continued investments in training, with integration with Hugging Face Optimum resulting in the acceleration of several Hugging Face models. New features include the ability to use native ORT ops in custom ops and call directly into a native or runtime operator without building a graph. The release also includes support for .NET 6 and the Multi-platform App UI (MAUI) and execution providers for specific platforms like the Neural Processing Unit on Android and Core ML on iOS. Performance improvements were made by reducing memory allocations during inferencing and eliminating unnecessary logging. Future improvements to enhance cache locality and thread pool utilization are planned.
v1.13 ONNX Runtime - Release Review
v1.13 ONNX Runtime - Release Review
Version 1.13 of the ONNX runtime was recently released with security patches, bug fixes, and performance enhancements. The update focuses on optimizing Transformer models for GPU quantization and adds support for direct ML execution providers that are device agnostic and support over 150 operators. Additionally, the release includes updates to the ORT mobile infrastructure for compatibility with new EPS, such as the XNN pack. The use of quantization to improve the performance of Transformer-based models is also discussed, with optimization of the CUDA execution provider to run the quantized BERT model and the use of quantized aware training to maximize accuracy while optimizing the ONNX runtime execution engine.