Skip to content
This repository has been archived by the owner on Feb 26, 2020. It is now read-only.
/ onnx-xla Public archive

XLA integration of Open Neural Network Exchange (ONNX)

Notifications You must be signed in to change notification settings

onnx/onnx-xla

Repository files navigation

onnx-xla

TODO:

  1. Fix bug in python setup.py develop (unable to copy module from build_ext)

  2. Experiment with group convolution implementations to improve performance.

  3. Use utility macros in onnx_xla and python_onnxifi to perform asserts

  4. Clean up onnx_xla/backend.cc macros

  5. Maybe move to one client (benchmark this?)

  6. Add support for half in python interface to ONNXIFI

  7. Add strided numpy array support ot python interface to ONNXIFI

  8. Add weight descriptor support to the python interface to ONNXIFI

  9. Benchmark two version of LRN(materializing square and not)

Steps to test:

  1. Run "python setup.py install" or "python setup.py develop"

  2. Start an XLA server with "./third_party/tensorflow/bazel-bin/tensorflow/compiler/xla/rpc/grpc_service_main_cpu --port=51000

  3. To the backends ability to run a simple IR graph with a Relu operator, "cd build && ./tests"

  4. To the backends ability to run a simple ModelProto graph with a Relu operator, "cd build && ./relu_model"

  5. To execute a test using the python wrapper of onnxifi, "python test.py"

  6. To run unit tests of node translations, execute a "python onnx_xla_test.py"

About

XLA integration of Open Neural Network Exchange (ONNX)

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published