http://dwnl3pr7ombq2abcycysnlqcl7rd6lqbavnir2swapbfyzrz4e7ae4id.onion/blog/what_model_to_use.md
A typical conversation might use 1,000-2,000 tokens, meaning costs can quickly accumulate with proprietary models. Also, new models using inference time compute are becoming increasingly available (currently only GPT-o1). These models are more expensive because they use more compute to generate each token.