Tag: Transformer
All the talks with the tag "Transformer".
Large Concept Models
Adhilsha AnsadPublished: at 07:00 PMThis talk will will explore the training objectives, segmentation techniques, and generation strategies of Large Concept Models (LCMs). LCMs are a novel class of models that leverage sentence-level tokenization to represent concepts, a higher abstraction than current tokens. We will also discuss the quantization of LCMs and their potential applications in various domains.
Input Combination Strategies for Multi-Source Transformer Decoder
Nalin KumarPublished: at 03:00 PMIn this talk, we will discuss the different input combination strategies for the encoder-decoder attention in the Transformer architecture.
Optimizing Transformer for Low-Resource Neural Machine Translation
Deepak KumarPublished: at 03:00 PMIn this talk, we will look at the effectiveness of the Transformer model under low-resource conditions and how the translation quality can be improved by optimizing the hyper-parameters.