Skip to content

Supported Models

twinny is a configurable extension, meaning many AI models are technically supported. However, not all models perform optimally with twinny in every scenario. This page lists models that have been tested and are recommended for specific features.

If you discover a model that works well with twinny but isn’t listed here, please open an issue or pull request to suggest its addition.

Chat Models

Most chat models trained for instruction-following should work with twinny. Here are some recommended examples:

Code Completion (Fill-in-the-Middle) Models

Only certain models are trained to support the fill-in-the-middle technique effectively. The following are examples of models recommended for this type of code completion. If you find another model that works well, please let us know.

Qwen2.5-coder Models

Codellama Models

Use the code suffixed versions of Codellama models.

Note: The 34b version of Codellama (e.g., codellama:34b-code) does not perform well for fill-in-the-middle tasks.

Deepseek Coder Models

Use the base versions of Deepseek Coder models.

Note: Models that are not base versions (e.g., instruct versions) do not work well for fill-in-the-middle.

Starcoder Models

Use the base versions of Starcoder models. The default Ollama tags for Starcoder usually point to the base versions.

Note: starcoder2 may sometimes fail to stop generating text appropriately. Lowering the temperature and increasing the repeat penalty settings can help mitigate this. For potentially better results with Starcoder2, consider using a specific size like starcoder2:7b.

Stablecode Models

Use the code suffixed versions of Stablecode models.

Codegemma Models

Use the code suffixed versions of Codegemma models.

Note: codegemma may also sometimes fail to stop generating text appropriately. Lowering the temperature and increasing the repeat penalty settings can help.