The recently released Neural Network Toolbox Converter for ONNX Model Format now allows one to export a trained Neural Network Toolbox™ deep learning network to the ONNX™ (Open Neural Network Exchange) model format. The ONNX model can then be imported into other deep learning frameworks, such as TensorFlow®, that support ONNX model import.
Alternatively, you could export via the MATLAB Compiler SDK.
Using the MATLAB Compiler SDK, you can save the trained network as a MAT file, and write a MATLAB function that loads the network from the file, performs the desired computation, and returns the Network's output.
You can then compile your MATLAB function into a shared library to be used in your C/C++, .NET, Java, or Python project.
You can find more information about MATLAB Compiler SDK in the following link:
Furthermore, the objects that MATLAB uses to represent Neural Networks are transparent, and you can therefore access all the information that describes your trained network.
For example, you will get an object of type SeriesNetwork, which is a trained Convolutional Neural Network. You can then see the weights and biases of the trained network:
Then, using for example caffe's MATLAB interface, you should be able to save a Convolutional Neural Network as a caffe model. The code for the MATLAB interface is in the following link:
and includes a classification demo that shows you how to use the interface.
Please note that the above code is not developed or supported by MathWorks Technical Support. If you have any questions about how to use the code, please contact the project's developers.