transformer models
a type of computer model that uses a mechanism called ‘attention’ to focus on different parts of the input at the same time. This helps the model learn how elements, such as words in a sentence, relate to each other, making it especially powerful for tasks like translation, summarization, and other kinds of pattern recognition. Language models that use this attention mechanism are in general referred to as large language models.