Skip to content

Releases: eaplatanios/tensorflow_scala

Updated to TensorFlow 2.3.1

08 Oct 19:13
Compare
Choose a tag to compare

Switched to using TensorFlow 2.3.1 that fixes multiple
security vulnerabilities of TensorFlow 2.3.0. Also:

  • Fixed a minor bug related to optimizer slot variables not being saved in TF checkpoints when tf.stopGradients is used.
  • Switched to supporting three native platforms: linux, windows, and darwin, where the last one does not include GPU support.

Fixed Precompiled TF Binaries and Added Some New Features

22 Nov 07:14
v0.4.1
1fc81b3
Compare
Choose a tag to compare

Fixed the precompiled TensorFlow binaries, and also added the following
new features:

  • io module:
    • Added support for a new TFRecordWriter.
  • ops module:
    • Added a new ops namespace, sparse, that includes all sparse ops.
    • Added support for sparse.reorder and sparse.merge.
    • Added support for parsing TF records.
    • data module:
      • Added support for Dataset.shuffleAndRepeat.
    • optimizers module:
      • Added support for the Adafactor optimizer.
      • Renamed SqrtDecay to RSqrtDecay which is more appropriate.
    • math module:
      • Added support for batchGather.
      • Added support for bitwise ops.
    • rnn module:
      • Simplified the attention mechanisms functionality so that it is
        now not required to tile memory tensors for beam search outside
        the beam search decoder.
    • Moved the seq2seq module to a separate repository (that of
      Symphony Machine Translation).

More Static Data Types

25 Oct 17:52
v0.4.0
7677354
Compare
Choose a tag to compare

This is a major release with a lot of new features related to static types for tensors and ops. The graph construction API is now statically-typed, thus enabling much better type safety than before.

Tensors and outputs are now statically-typed and the types used are the Scala types that correspond to the tensors' TensorFlow data types. For example:

val t1 = Tensor(0.5, 1) // The inferred type is Tensor[Double].
val t2 = Tensor(1, 2)   // The inferred type is Tensor[Int].
val t3 = t1 + t2        // The inferred type is Tensor[Double].
val t4 = t3.isNaN       // The inferred type is Tensor[Boolean].
val t5 = t3.any()       // Fails at compile-time because `any()` is only
                        // supported for Tensor[Boolean].

A similar situation now applies to Outputs. Ops are also typed and so is the auto-differentiation implementation.

This resulted in major simplifications in the data pipeline and the high level learn API. Datasets and dataset iterators do not "carry" T, V, D, and S types with them now, but rather just the type of the elements they contain/produce.

A new type trait called TF is also introduced that denotes supported Scala types in TensorFlow (e.g., TF[Int] and TF[Float]). Similarly, some more type traits are introduced to denote type constraints for various ops (e.g., IsIntOrUInt[Int], IsIntOrUInt[Long], IsFloatOrDouble[Float], etc.). These type traits are powered by a general implementation of union types for Scala.

Other new features include:

  • data module:
    • Added support for the mapAndBatch transformation.

Static Data Types and More

10 Oct 18:15
Compare
Choose a tag to compare

With this release we have finally added support for static data type
information for tensors (not for symbolic tensors yet though -- for now
we effectively have support for a statically-typed version of numpy
for Scala). This is an important milestone and contributes significantly
to type safety, which can help catch errors at compile time, rather than
runtime. For example:

val t1 = Tensor(0.5, 1) // The inferred type is Tensor[FLOAT64].
val t2 = Tensor(1, 2)   // The inferred type is Tensor[INT32].
val t3 = t1 + t2        // The inferred type is Tensor[FLOAT64].
val t4 = t3.isNaN       // The inferred type is Tensor[BOOLEAN].
val t5 = t3.any()       // Fails at compile-time because `any()` is only
                        // supported for Tensor[BOOLEAN].

Other new features include:

  • Improvements to the high-level learn API:
    • Layers can now provide and use their own parameter generator, and
      can also access the current training step
      (using Layer.currentStep).
    • Layers now support .map(...).
    • Added support for batch normalization.
  • Added support for tf.logSigmoid and tf.lrn.
  • Added support for the following new metrics:
    • Grouped precision.
    • Precision-at-k.
  • data module:
    • Added support for loading the extreme classification repository
      datasets (i.e., data.XCLoader).
    • Added support for randomly splitting datasets.

Minor Fix

25 Jun 23:54
v0.2.4
91de5e0
Compare
Choose a tag to compare

Fixed an issue with the packaged pre-compiled TensorFlow binaries that
affected Linux platforms.

Compatibility with TensorFlow 1.9-rc1

25 Jun 19:29
v0.2.3
6dab96a
Compare
Choose a tag to compare

Added compatibility with TensorFlow 1.9-rc1.

Pre-compiled Binaries Update

18 Jun 01:21
v0.2.2
11f11b4
Compare
Choose a tag to compare

In this release we have updated the precompiled TensorFlow binaries
distributed with this library.

Packaging Fix

17 Jun 18:05
v0.2.1
d5fa1b3
Compare
Choose a tag to compare

In this release we have fixed an issue related to the packaging and distributing of the pre-compiled TensorFlow shared libraries.

Updates

13 Jun 01:21
v0.2.0
8b8e58d
Compare
Choose a tag to compare

In this release we have:

  • Added support for incremental compilation.
  • Added support for Horovod.
  • Added support for timelines to allow for easy profiling of TensorFlow graphs.
  • Fixed a major memory leak (issue #87).
  • Updated the JNI bindings to be compatible with the TensorFlow 1.8.0 release.
  • Added support for obtaining the list of available devices from within Scala.
  • Fixed bugs for some control flow ops.
  • Added support for tf.cases.
  • Added support for the RMSProp optimizer, the lazy Adam optimizer, the AMSGrad optimizer, the lazy AMSGrad optimizer, and the YellowFin optimizer.
  • Added more learning rate decay schemes:
    • Cosine decay.
    • Cycle-linear 10x decay.
    • Square-root decay.
    • More warm-up decay schedules.
  • Added support for dataset interleave ops.
  • Fixed some bugs related to variable scopes and variable sharing.
  • Fixed some bugs related to functional ops.
  • Added support for some new image-related ops, under the namespace tf.image.
  • Improved consistency for the creation of initializer ops.
  • Added support for the tf.initializer op creation context.
  • Exposed part of the TensorArray API.
  • Exposed tf.Op.Builder in the public API.
  • Improvements to the learn API:
    • Refactored mode into an implicit argument.
    • Improved the evaluator hook.
    • Removed the layer creation context mechanism, to be refactored later. It was causing some issues due to bad design and unclear semantics. The plan is to implement this, in the near future, as wrapper creation context layers.
    • Improved the Model class.
    • Fixed a bug that was causing some issues related to inference hooks in the in-memory estimator.
    • Improved logging.
  • Added support for reading and writing numpy (i.e., .npy) files.
  • Added a logo. :)

Minor Fixes

31 Jan 06:55
Compare
Choose a tag to compare

This release fixes the following bugs:

  • Issue with the packaged pre-compiled TensorFlow binaries that
    affected Linux platforms.
  • Learn API bug where the shared name of input iterators was being
    set incorrectly.

I also switched to using CircleCI for continuous integration, instead
of TravisCI.