GPT-3
OpenAI, 2020. GPT-3 is a good historical anchor because its paper made the scale very visible: 175 billion parameters.
It is older than current frontier models, but it is still one of the clearest public examples of what “hundreds of billions of parameters” looks like in practice.