Tag: Transformer
All the talks with the tag "Transformer".
- Large Concept ModelsAdhilsha AnsadPublished: at 07:00 PM- This talk will will explore the training objectives, segmentation techniques, and generation strategies of Large Concept Models (LCMs). LCMs are a novel class of models that leverage sentence-level tokenization to represent concepts, a higher abstraction than current tokens. We will also discuss the quantization of LCMs and their potential applications in various domains. 
- Input Combination Strategies for Multi-Source Transformer DecoderNalin KumarPublished: at 03:00 PM- In this talk, we will discuss the different input combination strategies for the encoder-decoder attention in the Transformer architecture. 
- Optimizing Transformer for Low-Resource Neural Machine TranslationDeepak KumarPublished: at 03:00 PM- In this talk, we will look at the effectiveness of the Transformer model under low-resource conditions and how the translation quality can be improved by optimizing the hyper-parameters.