- Transformer Models are programmable. You can provide them with enough context (i.e., Memory), possibly from a data store (i.e., Virtual Memory). They then perform some processing and give you a response. For more info on prompt engineering see this
- Many practical NLP tasks can be cast as word prediction.
- Transformers are so effective because learning from human language is a cheat. Human language encodes human thinking, and transformers learn to exploit this.