onnx-scala-proto


License

License

Categories

Categories

Scala Languages
GroupId

GroupId

org.emergent-order
ArtifactId

ArtifactId

onnx-scala-proto_2.13
Last Version

Last Version

0.14.0
Release Date

Release Date

Type

Type

jar
Description

Description

onnx-scala-proto
onnx-scala-proto
Project URL

Project URL

https://github.com/EmergentOrder/onnx-scala
Project Organization

Project Organization

org.emergent-order
Source Code Management

Source Code Management

https://github.com/EmergentOrder/onnx-scala

Download onnx-scala-proto_2.13

How to add to project

<!-- https://jarcasting.com/artifacts/org.emergent-order/onnx-scala-proto_2.13/ -->
<dependency>
    <groupId>org.emergent-order</groupId>
    <artifactId>onnx-scala-proto_2.13</artifactId>
    <version>0.14.0</version>
</dependency>
// https://jarcasting.com/artifacts/org.emergent-order/onnx-scala-proto_2.13/
implementation 'org.emergent-order:onnx-scala-proto_2.13:0.14.0'
// https://jarcasting.com/artifacts/org.emergent-order/onnx-scala-proto_2.13/
implementation ("org.emergent-order:onnx-scala-proto_2.13:0.14.0")
'org.emergent-order:onnx-scala-proto_2.13:jar:0.14.0'
<dependency org="org.emergent-order" name="onnx-scala-proto_2.13" rev="0.14.0">
  <artifact name="onnx-scala-proto_2.13" type="jar" />
</dependency>
@Grapes(
@Grab(group='org.emergent-order', module='onnx-scala-proto_2.13', version='0.14.0')
)
libraryDependencies += "org.emergent-order" % "onnx-scala-proto_2.13" % "0.14.0"
[org.emergent-order/onnx-scala-proto_2.13 "0.14.0"]

Dependencies

compile (2)

Group / Artifact Type Version
org.scala-lang : scala-library jar 2.13.5
com.thesamet.scalapb : scalapb-runtime_2.13 jar 0.11.3

Project Modules

There are no modules declared in this project.


Build status Latest version

Getting Started

Add this to the build.sbt in your project:

libraryDependencies += "org.emergent-order" %% "onnx-scala-backends" % "0.13.0"

A short, recent talk I gave about the project: ONNX-Scala: Typeful, Functional Deep Learning / Dotty Meets an Open AI Standard

Full ONNX model inference - quick start

First, download the model file for SqueezeNet. You can use get_models.sh

Note that all code snippets are written in Scala 3 (Dotty).

First we create an "image" tensor composed entirely of pixel value 42:

import java.nio.file.{Files, Paths}
import org.emergentorder.onnx.Tensors._
import org.emergentorder.onnx.backends._
import org.emergentorder.compiletime._
import io.kjaer.compiletime._

val squeezenetBytes = Files.readAllBytes(Paths.get("squeezenet1.1.onnx"))
val squeezenet = new ORTModelBackend(squeezenetBytes)

val data = Array.fill(1*3*224*224){42f}

//In NCHW tensor image format
val shape =                    1     #:     3      #:    224    #: 224     #: SNil
val tensorShapeDenotation = "Batch" ##: "Channel" ##: "Height" ##: "Width" ##: TSNil

val tensorDenotation: String & Singleton = "Image"

val imageTens = Tensor(data,tensorDenotation,tensorShapeDenotation,shape)

//or as a shorthand if you aren't concerned with enforcing denotations
val imageTensDefaultDenotations = Tensor(data,shape)

Note that ONNX tensor content is in row-major order.

Next we run SqueezeNet image classification inference on it:

val out = squeezenet.fullModel[Float, 
                               "ImageNetClassification",
                               "Batch" ##: "Class" ##: TSNil,
                               1 #: 1000 #: SNil](Tuple(imageTens))
// val out:
//  Tensor[Float,("ImageNetClassification", 
//                "Batch" ##: "Class" ##: TSNil,
//                1 #: 1000 #: SNil)] = (Array(0.8230729,
// ...

//The output shape
out.shape
// val res0: Array[Int] = Array(1, 1000)


//The highest probability (predicted) class
out.data.indices.maxBy(out.data)
// val res1: Int = 418

Referring to the ImageNet 1000 class labels, we see that the predicted class is "ballpoint pen".

Based on a simple benchmark of 100000 iterations of SqueezeNet inference (run on my laptop), the run time is roughly on par (within 10% of) ONNX Runtime (via Python). The discrepancy can be accounted for by the overhead of shipping data between the JVM and native memory.

When using this API, we load the provided ONNX model file and pass it as-is to the underlying ONNX backend. This is the most performant execution mode, and is recommended for off-the-shelf models / performance-critical scenarios.

This full-model API is untyped in the inputs, so it can fail at runtime. This inevitable because we load models from disk at runtime. Feel free to wrap your calls into it in a facade with typed inputs.

Project Details

ONNX-Scala is cross-built against Scala JVM and Scala.js/JavaScript (for Scala 2.13 and Dotty/3.0)

Currently at ONNX 1.8.0 (Backward compatible to at least 1.2.0 for the full model API, 1.7.0 for the fine-grained API), ONNX Runtime 1.7.0.

Fine-grained API

A complete*, versioned, numerically generic, type-safe / typeful API to ONNX(Open Neural Network eXchange, an open format to represent deep learning and classical machine learning models), derived from the Protobuf definitions and the operator schemas (defined in C++).

We also provide implementations for each operator in terms of a generic core operator method to be implemented by the backend. For more details on the low-level fine-grained API see here

The preferred high-level fine-grained API, most suitable for the end user, is NDScala

* Up to roughly the intersection of supported ops in ONNX Runtime and ONNX.js

Training

Automatic differentiation to enable training is under consideration (ONNX currently provides facilities for training as a tech preview only).

Type-safe Tensors

Featuring type-level tensor and axis labels/denotations, which along with literal types for dimension sizes allow for tensor/axes/shape/data-typed tensors. Type constraints, as per the ONNX spec, are implemented at the operation level on inputs and outputs, using union types, match types and compiletime singleton ops (thanks to @MaximeKjaer for getting the latter into dotty). Using ONNX docs for dimension and type denotation, as well as the operators doc as a reference, and inspired by Nexus, Neurocat and Named Tensors.

Backend

There is one backend per Scala platform. For the JVM the backend is based on ONNX Runtime, via their official Java API. For Scala.js / JavaScript a backend based on ONNX.js is coming soon (blocked on new Scala.js bundler / ScalablyTyped converter releases for dotty support).

Supported ONNX input and output tensor data types:

  • Byte
  • Short
  • Int
  • Long
  • Float
  • Double
  • Boolean

Supported ONNX ops:

  • ONNX-Scala, Fine-grained API: 88/156 total

  • ONNX-Scala, Full model API: Same as below, depending on platform

  • ONNX Runtime: 145/156 total.

  • ONNX JS: 72/156 total.

See the ONNX backend scoreboard

Example execution

TODO: T5 example

Build / Publish

You'll need sbt.

To build and publish locally:

sbt publishLocal

or

sbt +publishLocal

to build against Scala 2.13 and Dotty/3.0, where possible.

Built With

Core

  • ONNX via ScalaPB - Open Neural Network Exchange / The missing bridge between Java and native C++ libraries (For access to Protobuf definitions, used in the fine-grained API to create ONNX models in memory to send to the backend)

  • Spire - Powerful new number types and numeric abstractions for Scala. (For support for unsigned ints, complex numbers and the Numeric type class in the core API)

  • Dotty - The Scala 3 compiler, also known as Dotty. (For union types (used here to express ONNX type constraints), match types, compiletime singleton ops, ...)

Backends

Inspiration

Scala

  • Neurocat - From neural networks to the Category of composable supervised learning algorithms in Scala with compile-time matrix checking based on singleton-types

  • Nexus - Experimental typesafe tensors & deep learning in Scala

  • Lantern - Machine learning framework prototype in Scala. The design of Lantern is built on two important and well-studied programming language concepts, delimited continuations (for automatic differentiation) and multi-stage programming (staging for short).

  • DeepLearning.scala - A simple library for creating complex neural networks

  • tf-dotty - Shape-safe TensorFlow in Dotty

Versions

Version
0.14.0
0.13.0
0.12.0