Between GPT-Neo and GPT-2, here are some factors to consider to help decide which is better for your use case:
1. Architecture and Size
- GPT-Neo is based on the architecture of GPT-3 and is available in sizes from 125M to 2.7B parameters. The smallest model (125M) may be a good balance for speed and capability on your hardware.
- GPT-2 has fewer options for parameters, from 117M to 1.5B, with the 117M model being very light. It’s tried and tested for general text generation tasks.