<aside> <img src="https://prod-files-secure.s3.us-west-2.amazonaws.com/03c9ca2b-b7d4-43c1-b15e-034f9d95f556/3a3b9b46-4383-47ea-9a26-f7315384e231/16.png" alt="https://prod-files-secure.s3.us-west-2.amazonaws.com/03c9ca2b-b7d4-43c1-b15e-034f9d95f556/3a3b9b46-4383-47ea-9a26-f7315384e231/16.png" width="40px" />

Please review the content below carefully. If it does not resolve your issue, we encourage you to explore our Education & Tutorials section before reaching out to Support. If you need us, we’re here to assist you and will do our best to help!

</aside>

Bolt.new inference uses Anthropic’s Sonnet 3.5 AI model. We purchase tokens from Anthropic, defined by them as: “the smallest individual units of a language model, and can correspond to words, subwords, characters, or even bytes (in the case of Unicode).” When users interact with Bolt, tokens are consumed in three primary ways: chat messages between the user and the LLM, the LLM writing code, and the LLM reading the existing code to capture any changes made by the user.

Our goal is for Bolt to use as few tokens as possible to accomplish each task. As such, the team is hard at work continuing to ship product changes that increase token efficiency.

Below are a number of tips you can currently implement to maximize token efficiency: