Local Providers
Best Practices Most settings — like model architecture and GPU offloading — can be adjusted via your LLM providers like LM Studio. However,max_tokens
and context_window
should be set via Open Interpreter. For local mode, smaller context windows will use less RAM, so we recommend trying a much shorter window (~1000) if it’s is failing or if it’s slow.
interpreter --local --max_tokens 1000 --context_window 3000
Make sure max_tokens
is less than context_window
.
Responses are generated using AI and may contain mistakes.
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4