Supported Models
twinny is a configurable extension, meaning many AI models are technically supported. However, not all models perform optimally with twinny in every scenario. This page lists models that have been tested and are recommended for specific features.
If you discover a model that works well with twinny but isn’t listed here, please open an issue or pull request to suggest its addition.
Chat Models
Most chat models trained for instruction-following should work with twinny. Here are some recommended examples:
Code Completion (Fill-in-the-Middle) Models
Only certain models are trained to support the fill-in-the-middle technique effectively. The following are examples of models recommended for this type of code completion. If you find another model that works well, please let us know.
Qwen2.5-coder Models
Codellama Models
Use the code
suffixed versions of Codellama models.
Note: The 34b
version of Codellama (e.g., codellama:34b-code
) does not perform well for fill-in-the-middle tasks.
Deepseek Coder Models
Use the base
versions of Deepseek Coder models.
Note: Models that are not base
versions (e.g., instruct
versions) do not work well for fill-in-the-middle.
Starcoder Models
Use the base
versions of Starcoder models. The default Ollama tags for Starcoder usually point to the base
versions.
starcoder
starcoder2
(e.g.,starcoder2:7b
)
Note: starcoder2
may sometimes fail to stop generating text appropriately. Lowering the temperature and increasing the repeat penalty settings can help mitigate this. For potentially better results with Starcoder2, consider using a specific size like starcoder2:7b
.
Stablecode Models
Use the code
suffixed versions of Stablecode models.
Codegemma Models
Use the code
suffixed versions of Codegemma models.
Note: codegemma
may also sometimes fail to stop generating text appropriately. Lowering the temperature and increasing the repeat penalty settings can help.