The transformer is a deep learning architecture that was developed by researchers at Google and is based on the multi-head attention mechanism, which was proposed in the 2017 paper "Attention Is All You Need". From Wikipedia
The new model integrates hybrid architecture for improved speed and reasoning, aiming to rival DeepSeek's AI dominance.