Too many basic ops which TF has such as tf.einsum(…) missing in DL4J, is it possible to encapsulate TF c/c++ interface to have such ops?
@SidneyLann Many ops aren’t needed for inference which is the main target. I am evaluating doing something like this by allowing dynamic wrapping of other ops. If you want to contribute something, feel free to open an issue to discuss it, otherwise I’ll try to introduce something like that. As it stands, a short term goal I have is to add all ops that are both in TF and defined in onnx: onnx/Operators.md at master · onnx/onnx · GitHub
There’s actually not that many missing. It’s a fair point though and I’ll keep that in mind when we’re doing the next release.
DL4J to me is for building model and trainning, not only inference. I want to port GAT from py and some basic ops not found in DL4J like tf.einsum etc. In fact, I had never used TF and no the py skill, just have skill reading py source.
Up to now, DL4J has all featrues I want except GAT.
einsum is a pretty complex op though, and unless you need dynamic mapping, you should be able to implement whatever underlying computation you need directly.
Yes. So I think if it is possible to implement nd4j=libnd4j+tf c++?
What I’m saying is that you can implement what einsum is doing manually, because it is usually doing one of the things that is actually implemented in ND4J.
Mixing tf and nd4j isn’t something you can do.
attn_for_self = tf.einsum("…NHI , IHO -> …NHO", x, self.attn_kernel_self)
What SameDiff ops in ND4J should I use then? Or just combine INDArray methods?