A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from https://docs.openinterpreter.com/language-models/local-models/best-practices below:

Best Practices - Open Interpreter

Local Providers

Best Practices

Most settings — like model architecture and GPU offloading — can be adjusted via your LLM providers like LM Studio. However, max_tokens and context_window should be set via Open Interpreter. For local mode, smaller context windows will use less RAM, so we recommend trying a much shorter window (~1000) if it’s is failing or if it’s slow.
interpreter --local --max_tokens 1000 --context_window 3000

Make sure max_tokens is less than context_window.

Responses are generated using AI and may contain mistakes.


RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4