Hey Brad Larson,
I very much love S4TF project. But I find that sometimes reading documentation is difficult in comparison to that of PyTorch. For instance, maybe some newcomer to S4TF might not know/remember the formula for a given function, say `categoricalHingeLoss(predicted:expected:)`, then it becomes a difficult & lengthy process to Google it and then come back to continue learning and using it.
I believe it will be amazing if equations describing the functions (say, categoricalHingeLoss(predicted:expected:)), structures (say, Conv[x]D) ], and more alongside their usage examples are provided in documentation. Basically, please add three more things to S4TF documentation to make it easy to use:
1. Long description of ML concept. Link:
https://pytorch.org/docs/stable/nn.html#transformer-layers
- For instance,
A transformer model. User is able to modify the attributes as needed. The architecture is based on the paper “Attention Is All You Need”. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems, pages 6000-6010. Users can build the BERT(
https://arxiv.org/abs/1810.04805) model with corresponding parameters.
2. Equations describing machine learning concepts.
3. Example usage code for each concept.
Currently, the way documentation is designed sometimes hinders the process of using S4TF smoothly. Since Swift is a relatively new language for machine learning it needs to be cool from every viewpoint so as to attract new users. I think these important modifications (maybe more) to documentation will welcome a wide range of people into trying and using Swift for TensorFlow. Please take some action, it will be really helpful to me too!
- Rahul Bhalley