Transformer/Attention-NLP

Hello,

Can you please point me to end to end examples that would use transformer layer and multihead attention with dl4j.

Thanks
Ravi.

Trying to build following example: https://keras.io/examples/nlp/text_classification_with_transformer/ using dl4j

All attention models can be found here:

I would also suggest running keras import to see if that fits your use case:
https://deeplearning4j.konduit.ai/keras-import/overview

Thanks for the response - I am trying to build attention/transformer natively using java only. I couldn’t find any example that would show how to do that.
Are you suggesting/encouraging that attention models be implemented in keras, and loaded as keras layer in dl4j something like a hybrid approach?