Open fchollet opened 1 year ago
Various (undocumented!) backend functions missing
I checked out the issue for converting keras.io examples to keras_core
examples. One important functionality that is missing from the backend is CTC
. Widely used but diff frameworks implement it in different ways. If we are expecting people to use keras_core
for writing framework agnostic code, we need to implement it in the backend.
Hello, I wanted to ask about the plans to add a KerasRaggedTensor to Keras Core. I would be very interested to have this feature, just to feed the Tensors of shape (batch, None, C, F, ...) to a keras model. I understand that keras backend ops can not support ragged operations across platforms but simply to pass a ragged Tensor to layers as input and output and extract splits and values would be super helpful.
I would be very interested to have this feature, just to feed the Tensors of shape (batch, None, C, F, ...) to a keras model
We might support RaggedTensor
in the future with the TF backend specifically (such a feature does not exist with other frameworks).
However if you're just looking to have a dynamic data dimension, then you can do it just by:
(batch, A, ...)
, (batch, B, ...)
, etc. and creating batches out of the buckets, so that each batch is rectangularIn general it is possible to handle any workflow using rectangular tensors. RaggedTensors are a convenience but not a blocker.
I would be very interested to have this feature, just to feed the Tensors of shape (batch, None, C, F, ...) to a keras model
We might support
RaggedTensor
in the future with the TF backend specifically (such a feature does not exist with other frameworks).However if you're just looking to have a dynamic data dimension, then you can do it just by:
- Bucketing your samples into buckets of shape
(batch, A, ...)
,(batch, B, ...)
, etc. and creating batches out of the buckets, so that each batch is rectangular- Padding/truncating your samples to a shared shape
- Some combination of the two
In general it is possible to handle any workflow using rectangular tensors. RaggedTensors are a convenience but not a blocker.
No idea how it is UX or performance wise, but torch recently added https://pytorch.org/torchrec/torchrec.sparse.html
“jagged tensor” instead of “ragged”
A small number of items are likely to affect you ..., TF SavedModel support changes,
For backend agnostic, this may reasonable change (but quite big change). What about the .h5
format? Recently while trying to save keras model, I got the following, is it intended? What the diff of .weights.h5
or only .h5
? I think I could do (.h5
) earlier.
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
[<ipython-input-7-694b2dd940b3>](https://localhost:8080/#) in <cell line: 14>()
12 model.compile(optimizer='adam', loss='mse')
13 _ = model(np.random.rand(96, 96))
---> 14 model.save_weights('model.h5')
1 frames
[/usr/local/lib/python3.10/dist-packages/keras/src/models/model.py](https://localhost:8080/#) in save_weights(self, filepath, overwrite)
371 """
372 if not str(filepath).endswith(".weights.h5"):
--> 373 raise ValueError(
374 "The filename must end in `.weights.h5`. "
375 f"Received: filepath={filepath}"
ValueError: The filename must end in `.weights.h5`. Received: filepath=model.h5
In the built-in model, the .h5
for weight only looks still valid (though it's from tf.keras it runs w/o raising above error).
keras.applications.EfficientNetV2B0(
include_top=True,
weights="imagenet",
input_tensor=None,
input_shape=None,
pooling=None,
classes=1000,
classifier_activation="softmax",
include_preprocessing=True,
)
output
Downloading data from https://storage.googleapis.com/tensorflow/keras-applications/efficientnet_v2/efficientnetv2-b0.h5
29403144/29403144 ━━━━━━━━━━━━━━━━━━━━ 0s 0us/step
<Functional name=efficientnetv2-b0, built=True>
I was told over at tensorflow that Keras 3 will not support nested input dictionaries. While I don't understand why that support was removed, that's definitely an incompatibility that's worth mentioning here.
compat.v1
APIs are gone (deprecated in 2019).
Hello, we are using tf.compat.v1.keras
in our project, and it works in TensorFlow 2.15. Does it mean we need to change our codes for TensorFlow 2.16?
Update: I found yes. Use tf-keras instead.
Is there any particular reason why AlphaDropout layer is removed? Is there an alternative we can use? I have some code examplesi am trying to convert to v3 and they use tf.keras.layers.AlphaDroput layer
Is there any particular reason why AlphaDropout layer is removed? Is there an alternative we can use? I have some code examplesi am trying to convert to v3 and they use tf.keras.layers.AlphaDroput layer
Same question here.
SELU needs AlphaDropout.
The reset_states
method of a model disappeared, and now it is only available on layers. Will this be permanent? This was a nice way of resetting all internal states of all RNN layers in a model.
@fchollet May I ask what the consideration is for "Layer names and variable names can no longer contain the / character."? I have to make many efforts to convert my old checkpoints and weights for keras-3 compatibility.
UPDATE:
The current temporary workaround for me is to wrap the keras.layer
and keras.model
to replace /
and then wrap the weights reader (e.g., h5) to replace '/` in weights as well.
There is no more method compute_output_signature
in Layer
. Now it is compute_output_spec
The
reset_states
method of a model disappeared, and now it is only available on layers. Will this be permanent? This was a nice way of resetting all internal states of all RNN layers in a model.
I suspect something may have changed re dynamic binding of methods or some other horrible pythonic thing.If I create a Sequential model:
Sequential().reset_states
<bound method Sequential.reset_states of <Sequential name=sequential_6, built=False>>
And when I subclass Sequential, I cannot call the parent method:
class Sequential(keras.models.Sequential):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
def reset_states(self):
print(cs("Sequential::reset_states", "cyan"))
super().reset_states()
'super' object has no attribute 'reset_states'
As a side note, please could these sorts of changes be included in formal API documentation? It seems most of the documentation on the Keras site is in the style of small tutorial examples, and a github issue to list breaking changes isn't the best way to navigate what needs to be done, especially when dealing with a scripting language and there is no compiler to help us out. Being able to navigate something like Doxygen and including breaking changes there would be a much better way of seeing hat has changed.
Thanks
Additionally, it seems that RNNs don't support Cells with multiple inputs anymore. Really sad. The "sequences" parameter (note plural) in the rnn.py
file assumes actually only one "sequence" (singular).
While trying to run TCN with Keras 3 https://github.com/philipperemy/keras-tcn/issues/256 (issue opened by another person) I found, that while Sequential model in earlier versions allows to get/set Weights in build method, it's not the case with Keras 3 (there is error that numpy is not available ("numpy() is only available when eager execution is enabled"), while previously this was not a case, also functional model somehow has no this problem) see example in first comment https://github.com/tensorflow/addons/issues/2869 (the error is with tensorflow addons master branch, which is not completely ported to Keras 3, but still demonstrates a problem)
additionally: support for causal padding is removed from SeparableConv1D in Keras 3, so it's more difficult to add SeparableConv1D to TCN
Hello, In undocumented backend function of tf_keras we had .set_value() and get_value() will those be added to keras 3 in future? Thankyou.
The list of behavior changes between the old tf.keras and Keras 3 with TF backend is missing the redefinition of the hard_sigmoid
activation function.
Old version: https://github.com/keras-team/keras/blob/v2.15.0/keras/backend.py#L5924 New version: https://github.com/keras-team/keras/blob/v3.0.0/keras/backend/tensorflow/nn.py#L53
I find it worrying that such a breaking change has been made without mentioning it anywhere. The only mention of this change I could find was this issue on the TF GitHub pages, which noted that the documentation for hard_sigmoid
was incorrect.
The problem is that this change silently alters the output of every model that uses hard_sigmoid
trained with old versions of tf.keras when used in TF >= 2.16, but nothing in the list of changes or migration guide indicates this.
Keras 3 is a major new release. It features a number of cleanups and modernizations of Keras which leads to number of breaking changes compared to Keras 2.
The list below is exhaustive to the best of our knowledge.
A small number of items are likely to affect you (
jit_compile
default value change, TF SavedModel support changes, usage oftf.Variable
as layer attributes). The majority are very niche. All APIs that were removed were dropped due to extremely low usage.Behavior differences between old
tf.keras
and Keras 3 (with TF backend)compat.v1
APIs are gone (deprecated in 2019). In the case ofexperimental
APIs, usually those are APIs that have already moved to a permanent namespace long ago (e.g. the contents oftf.keras.layers.experimental.preprocessing
is now atkeras.layers
, since 2021), so just update the import path to the up-to-date location.jit_compile=True
by default -- this might not work with all TF ops, so with some custom models/layers you might have setjit_compile=False
if you see an XLA related error.model.save()
is no longer supported (note: you can usetf.save_model.save(model)
instead)keras.models.load_model()
is no longer supported (note: you can usekeras.layers.TFSMLayer(filepath, call_endpoint="serving_default")
to reload any TF SavedModel as a Keras layer)Model()
can no longer be passed deeply nested inputs/outputs (nested more than 1 level deep, e.g. lists of lists of tensors)tf.keras
, TF autograph is enabled by default on thecall()
method of custom layers. In Keras 3, it is not. This means you may have to usecond
ops if you're using control flow, or alternatively you can decorate yourcall()
method with@tf.function
.keras.ops
.evaluate()
method does not return individual output losses anymore -> use themetrics
argument in compile to track them/
character.RaggedTensor
support. We may add it back later.output_a
andoutput_b
, oldtf.keras
adds<output_a>_loss
,<output_b>_loss
and so on to metrics. Keras 3 doesn't add them to metrics and needs to be done them to the output metrics by explicitly providing them inmetrics
list of individual outputs.tf.Variable
objects assigned to aLayer
are not tracked as part ofweights
. Fix: useself.add_weight()
method or use akeras.Variable
instead.None
entries are not allowed as part of nested (e.g. list/tuples) tensor arguments inLayer.call()
, nor as part ofcall()
return values.add_loss()
is removed (you can still useadd_loss()
inside thecall()
method of a layer/model).Layer
attributesmetrics
,dynamic
are removedconstants
arg in RNN layers is removed (remnant of Theano, ~0 usage)time_major
arg in RNN layers is removed (~0 usage)reset_metrics
argument frommodel.*_on_batch
methods (~0 usage)RadialConstraint
constraint object is removed (~0 usage)Present in Keras 3 standalone but will work when accessing Keras 3 via the new
tf.keras
backend.random_normal
AlphaDropout
layer is removedThresholdedReLU
layer is removed (subsumed byReLU
)RandomHeight
/RandomWidth
layers are removed (better useRandomZoom
)