+ Configure Ollama instances for distributed processing +
++ Discover and select models from your Ollama instances + {hasCache && lastDiscoveryTime && ( + + (Cached {new Date(lastDiscoveryTime).toLocaleTimeString()}) + + )} +
+{error}
+ ++ {discoveryProgress || `Scanning ${enabledInstanceUrls.length} Ollama instances...`} +
+No models found
++ {models.length === 0 + ? "Try refreshing to discover models from your Ollama instances" + : "Adjust your filters to see more models" + } +
++ {model.description} +
+ )} + + {/* Performance Metrics - flexible layout */} ++ Choose the best model for your needs ({modelType} models from {selectedInstanceUrl?.replace('http://', '') || 'all hosts'}) +
+No models found
+ +@@ -53,49 +511,457 @@ export const RAGSettings = ({ knowledge retrieval.
- {/* Provider Selection Row */} -Configure separate Ollama instances for LLM and embedding models
+For chat completions and text generation
+For generating text embeddings and vector search
+| Configuration | +LLM Instance | +Embedding Instance | +
|---|---|---|
| Instance Name | ++ {llmInstanceConfig.name || Not configured} + | ++ {embeddingInstanceConfig.name || Not configured} + | +
| Status | ++ + {llmStatus.checking ? "Checking..." : llmStatus.online ? `Online (${llmStatus.responseTime}ms)` : "Offline"} + + | ++ + {embeddingStatus.checking ? "Checking..." : embeddingStatus.online ? `Online (${embeddingStatus.responseTime}ms)` : "Offline"} + + | +
| Selected Model | ++ {getDisplayedChatModel(ragSettings) || No model selected} + | ++ {getDisplayedEmbeddingModel(ragSettings) || No model selected} + | +
| Available Models | +
+ {ollamaMetrics.loading ? (
+
+
+ )}
+ {ollamaMetrics.llmInstanceModels.total} Total Models
+ {ollamaMetrics.llmInstanceModels.total > 0 && (
+
+
+ {ollamaMetrics.llmInstanceModels.chat} Chat
+
+
+ {ollamaMetrics.llmInstanceModels.embedding} Embedding
+
+
+ )}
+ |
+
+ {ollamaMetrics.loading ? (
+
+
+ )}
+ {ollamaMetrics.embeddingInstanceModels.total} Total Models
+ {ollamaMetrics.embeddingInstanceModels.total > 0 && (
+
+
+ {ollamaMetrics.embeddingInstanceModels.chat} Chat
+
+
+ {ollamaMetrics.embeddingInstanceModels.embedding} Embedding
+
+
+ )}
+ |
+
+ Configure your Anthropic API key in the credentials section to use Claude models. +
++ Groq provides fast inference with Llama, Mixtral, and Gemma models. +
+