๐Ÿง  Thinking Filters

Simple filters for supporting models that use chain-of-thought reasoning capabilities. These filters rewrite output to separate reasoning/thought process from the final output.

What is Chain-of-Thought?

Chain-of-Thought prompting is an experimental way for large language models to reason about problems better. Because LLMs are fundamentally text prediction machines (that is, they predict what text will follow the text given to them), they often generate all kinds of strange yet confident answers. Chain-of-Thought reasoning aims to reduce these issues by training the model to output text for reasoning about user queries.

So while the LLM is still fundamentally generating predicted text based on input, that generated text is now designed to follow a reasoning path. This leads to notable reduction in hallucinations and calculation errors. It's not perfect, but it's better than nothing.

The biggest downside of Chain-of-Thought prompting is the significant increase in token output, resulting in slower response times and lots of extra text. It becomes difficult to focus on the essential information: the LLM's final answer. These filters mitigate this by providing a clear distinction between the model's thought process and its conclusion.

Filters and Supported Models

These are the current filters and what models they support:

โŒ› Filter: Artificium Thinking Filter ๐Ÿค– Supported Model: Artificium Llama 3.1 8b

โŒ› Filter: Collapsible Thought Filter ๐Ÿค– Supported Model: Reflection 70b

Configuration

Documentation is separated by filter.

Collapsible Thought Filter

Has 4 settings besides filter priority:

Artificium Thinking Filter

Has 3 settings besides filter priority:

Usage

These filters are meant to make output from Chain-of-Thought LLMs more readable. They are also designed specifically for certain models and their derivatives. Everything should work out of the box.

The "Use Thoughts as Context" setting is disabled by default because it will send a lot more tokens to the LLM, resulting in longer processing and fuller contexts. Enabling it might get more accurate answers over longer conversations.

โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—โ€—

โคด๏ธ [/projects/open-webui-filters] ๐Ÿ  Home