Skip to content

Quick start

Prerequisites

Before you start using twinny you need to have access to an inference provider. An inference provider is a local or cloud hosted server that runs the AI models.

The recommended way to do this is to use Ollama. Ollama makes it easy to run your models locally and exposes them as an OpenAI compatible API. Performance will depend on your hardware and chosen model, see Ollama’s documentation for more information.

Installing the extension

  1. Install the Visual Studio Code extension here or for VSCodium here.

Installing Ollama as an inference provider

  1. Visit Install Ollama and follow the instructions to install Ollama on your machine.
  2. Choose a model from the list of models available on Ollama. Two recommended models to get started are codellama:7b-instruct for chat and codellama:7b-code for fill-in-middle. See the Supported models page for more options.
Terminal window
ollama run codellama:7b-instruct
ollama run codellama:7b-code

Once both the extension and Ollama are installed you can start using twinny.

  1. Open VS code (if already open a restart might be needed after install) and look for the twinny icon in the side panel.

You should also see the 🤖 icon indicating that twinny is ready to use. The icon will change to a spinner when twinny is making a call to the inference provider.