Home Arrow Icon Knowledge base Arrow Icon Global Arrow Icon What are the main reasons for restricting the API to 32K tokens despite the model supporting 128K


What are the main reasons for restricting the API to 32K tokens despite the model supporting 128K


The restriction of the API to a maximum of 32,000 tokens, despite the model's capability to support up to 128,000 tokens, can be attributed to several key factors:

1. Performance and Stability**

Limiting the token count helps maintain the performance and stability of the API. Higher token limits can lead to increased computational demands, which may result in slower response times and potential instability during high-traffic periods. By capping the token limit at 32K, providers can ensure more consistent performance across different usage scenarios[1].

2. User Experience**

A lower token limit encourages users to craft more concise and focused queries. This can enhance the overall interaction quality by promoting clearer communication between users and the model. When users are limited to fewer tokens, they are more likely to engage in iterative dialogue rather than overwhelming the model with excessively lengthy inputs[3].

3. Resource Management**

Operating within a defined token limit allows for better resource management on the server side. Each request requires memory and processing power, and by restricting token usage, service providers can optimize their infrastructure to handle more simultaneous requests without degrading service quality[2].

4. Cost Considerations**

Higher token limits can lead to increased operational costs for service providers due to greater resource consumption. By maintaining a 32K limit, providers can manage costs more effectively while still offering a robust service that meets the needs of most users[6].

5. Technical Constraints**

There may also be technical constraints related to how models are deployed or configured in specific environments. For instance, some implementations might not fully support higher token limits due to limitations in their architecture or deployment settings[4][5].

In summary, while the model itself has the potential for a larger context window, practical considerations regarding performance, user experience, resource management, cost, and technical constraints drive the decision to restrict the API to 32K tokens.

Citations:
[1] https://www.linkedin.com/pulse/gpt-4-continues-lead-googles-32k-models-meet-match-token-ankit-pareek
[2] https://learn.microsoft.com/en-au/answers/questions/2121835/gpt-4o-has-a-35k-input-token-limit
[3] https://devblogs.microsoft.com/surface-duo/android-openai-chatgpt-15/
[4] https://learn.microsoft.com/en-us/answers/questions/2116223/azure-openai-token-limit-issue
[5] https://community.zapier.com/troubleshooting-99/chatgpt-error-400-max-token-is-too-large-32768-this-model-supports-at-most-4096-completion-tokens-39804
[6] https://themeisle.com/blog/chatgpt-api-cost/
[7] https://community.openai.com/t/anyone-can-explain-me-why-i-cannot-set-max-token-to-32k/951948
[8] https://www.reddit.com/r/OpenAI/comments/1h7jm52/one_thing_that_openai_shouldve_highlighted_more/
[9] https://github.com/danny-avila/LibreChat/discussions/1077
[10] https://www.googlecloudcommunity.com/gc/AI-ML/Gemini-1-0-Pro-tekon-count-not-32K/m-p/719426