News

ML.NET Now Works on ARM Devices and Blazor WebAssembly

ML.NET now works on ARM64 and Apple M1 devices, and on Blazor WebAssembly, with some limitations for each.

Microsoft regularly updates ML.NET, an open source, cross-platform machine learning (ML) framework for .NET devs, along with the integral Model Builder component that features a simple visual interface for building, training and deploying custom ML models without requiring deep expertise.

In the June 2021 update, the new ARM64 and Blazor WebAssembly functionality was unveiled.

"You can now perform training and inferencing with ML.NET on ARM64 and Apple M1 (in addition to Linux and macOS) devices which enables platform support for mobile and embedded devices as well as ARM-based servers," Microsoft announced.

Training and Inferencing on a Pinebook Pro Laptop Running Manjaro ARM Linux Distribution (source: Microsoft).

Limitations that could throw a DLL not found exception when training and inferencing on ARM include:

  • Symbolic SGD, TensorFlow, OLS, TimeSeries SSA, TimeSeries SrCNN and ONNX are not currently supported for training or inferencing.
  • LightGBM is currently supported for inferencing, but not training.
  • You can add LightGBM and ONNX support by compiling them for ARM, but they don’t provide pre-compiled binaries for ARM/ARM64.

For Blazor WebAssembly -- the client-side component of Blazor, which allows for C#-based web development instead of JavaScript -- some training and inferencing is possible on .NET 6, currently in preview. It has the same limitations as those listed for ARM above, plus two more:

  • You must currently set the EnableMLUnsupportedPlatformTargetCheck flag to false to install in Blazor.
  • LDA and Matrix Factorization are not supported.

According to the release notes for the new update (ML.NET v1.5.5), other new features include:

  • New API allowing confidence parameter to be a double.(#5623) . A new API has been added to accept double type for the confidence level. This helps when you need to have higher precision than an int will allow for.
  • Support to export ValueMapping estimator to ONNX was added (#5577)
  • New API to treat TensorFlow output as batched/not-batched (#5634) A new API has been added so you can specify if the output from TensorFlow is batched or not.

The release notes also detail many bug fixes and documentation updates.

For the Model Builder component specifically, features previously presented in preview are now publicly available, including:

  • Config-based training with generated code-behind files
  • Restructured Advanced Data Options
  • Redesigned Consume step

AutoML, used to automate the time-consuming, iterative tasks of machine learning model development, has also been improved in cooperation with Microsoft Research. Generally, that partnership is expected to provide these benefits:

  • Enabling AutoML support for all ML.NET scenarios
  • Allowing more precise control over the hyperparameter search space
  • Enabling more training environments, including local, Azure, and on-prem distributed training
  • Opening up future collaborations on advanced ML tech, like Network Architecture Search (NAS)

Among the first of specific concrete improvements are better benchmark tests reflecting an increased number of models that are explored for different train times, and other metrics.

The announcement post also includes results of a survey conducted by the dev team, revealing -- among many other things -- that the biggest blockers/pain points/challenges reported by respondents for using ML.NET include:

  • Small ML.NET community
  • Docs and samples (quantity, quality, real world)
  • Insufficient deep learning support
  • Specific ML scenario or algorithm not supported by ML.NET
  • Afraid Microsoft will abandon it

The post details how Microsoft is planning to address those.

About the Author

David Ramel is an editor and writer at Converge 360.

comments powered by Disqus

Featured

  • Hands On: New VS Code Insiders Build Creates Web Page from Image in Seconds

    New Vision support with GitHub Copilot in the latest Visual Studio Code Insiders build takes a user-supplied mockup image and creates a web page from it in seconds, handling all the HTML and CSS.

  • Naive Bayes Regression Using C#

    Dr. James McCaffrey from Microsoft Research presents a complete end-to-end demonstration of the naive Bayes regression technique, where the goal is to predict a single numeric value. Compared to other machine learning regression techniques, naive Bayes regression is usually less accurate, but is simple, easy to implement and customize, works on both large and small datasets, is highly interpretable, and doesn't require tuning any hyperparameters.

  • VS Code Copilot Previews New GPT-4o AI Code Completion Model

    The 4o upgrade includes additional training on more than 275,000 high-quality public repositories in over 30 popular programming languages, said Microsoft-owned GitHub, which created the original "AI pair programmer" years ago.

  • Microsoft's Rust Embrace Continues with Azure SDK Beta

    "Rust's strong type system and ownership model help prevent common programming errors such as null pointer dereferencing and buffer overflows, leading to more secure and stable code."

  • Xcode IDE from Microsoft Archrival Apple Gets Copilot AI

    Just after expanding the reach of its Copilot AI coding assistant to the open-source Eclipse IDE, Microsoft showcased how it's going even further, providing details about a preview version for the Xcode IDE from archrival Apple.

Subscribe on YouTube

Upcoming Training Events