A curated list of the latest breakthroughs in AI by release date with a clear video explanation, link to a more in-depth article, and code. [work in progress]
A Non-Autoregressive Transformer based TTS, supporting a family of SOTA transformers with supervised and unsupervised duration modelings. This project grows with the research community, aiming to achieve the ultimate TTS.
Thanks for your great work!
Could you please share the influence of the batch size and the number of GPUs?
Also how to choose a suitable learning rate and batch size if the available GPUs is not enough.
Thank you!
Absolutely amazing SOTA Google Colab (Jupyter) Notebooks for creating/training SOTA Music AI models and for generating music with Transformer technology (Google XLNet/Transformer-XL)
Pytorch Implementation of Deep neural network for traffic sign recognition systems: An analysis of spatial transformers and stochastic optimisation methods
Thanks for your great work!
Could you please share the influence of the batch size and the number of GPUs?
Also how to choose a suitable learning rate and batch size if the available GPUs is not enough.
Thank you!