forked from open-webui/open-webui
		
	main #2
					 15 changed files with 526 additions and 207 deletions
				
			
		
							
								
								
									
										10
									
								
								.github/workflows/build-release.yml
									
										
									
									
										vendored
									
									
								
							
							
						
						
									
										10
									
								
								.github/workflows/build-release.yml
									
										
									
									
										vendored
									
									
								
							|  | @ -29,11 +29,11 @@ jobs: | ||||||
|     - name: Extract latest CHANGELOG entry |     - name: Extract latest CHANGELOG entry | ||||||
|       id: changelog |       id: changelog | ||||||
|       run: | |       run: | | ||||||
|         CHANGELOG_CONTENT=$(awk '/^## \[/{n++} n==1' CHANGELOG.md) |         CHANGELOG_CONTENT=$(awk 'BEGIN {print_section=0;} /^## \[/ {if (print_section == 0) {print_section=1;} else {exit;}} print_section {print;}' CHANGELOG.md) | ||||||
|         echo "CHANGELOG_CONTENT<<EOF"  |         CHANGELOG_ESCAPED=$(echo "$CHANGELOG_CONTENT" | sed ':a;N;$!ba;s/\n/%0A/g') | ||||||
|         echo "$CHANGELOG_CONTENT" |         echo "Extracted latest release notes from CHANGELOG.md:"  | ||||||
|         echo "EOF"  |         echo -e "$CHANGELOG_CONTENT"  | ||||||
|         echo "::set-output name=content::${CHANGELOG_CONTENT}" |         echo "::set-output name=content::$CHANGELOG_ESCAPED" | ||||||
| 
 | 
 | ||||||
|     - name: Create GitHub release |     - name: Create GitHub release | ||||||
|       uses: actions/github-script@v5 |       uses: actions/github-script@v5 | ||||||
|  |  | ||||||
|  | @ -41,7 +41,7 @@ ENV WHISPER_MODEL_DIR="/app/backend/data/cache/whisper/models" | ||||||
| # for better persormance and multilangauge support use "intfloat/multilingual-e5-large" (~2.5GB) or "intfloat/multilingual-e5-base" (~1.5GB) | # for better persormance and multilangauge support use "intfloat/multilingual-e5-large" (~2.5GB) or "intfloat/multilingual-e5-base" (~1.5GB) | ||||||
| # IMPORTANT: If you change the default model (all-MiniLM-L6-v2) and vice versa, you aren't able to use RAG Chat with your previous documents loaded in the WebUI! You need to re-embed them. | # IMPORTANT: If you change the default model (all-MiniLM-L6-v2) and vice versa, you aren't able to use RAG Chat with your previous documents loaded in the WebUI! You need to re-embed them. | ||||||
| ENV RAG_EMBEDDING_MODEL="all-MiniLM-L6-v2" | ENV RAG_EMBEDDING_MODEL="all-MiniLM-L6-v2" | ||||||
| # device type for whisper tts and ebbeding models - "cpu" (default), "cuda" (nvidia gpu and CUDA required) or "mps" (apple silicon) - choosing this right can lead to better performance | # device type for whisper tts and embbeding models - "cpu" (default), "cuda" (nvidia gpu and CUDA required) or "mps" (apple silicon) - choosing this right can lead to better performance | ||||||
| ENV RAG_EMBEDDING_MODEL_DEVICE_TYPE="cpu" | ENV RAG_EMBEDDING_MODEL_DEVICE_TYPE="cpu" | ||||||
| ENV RAG_EMBEDDING_MODEL_DIR="/app/backend/data/cache/embedding/models" | ENV RAG_EMBEDDING_MODEL_DIR="/app/backend/data/cache/embedding/models" | ||||||
| ENV SENTENCE_TRANSFORMERS_HOME $RAG_EMBEDDING_MODEL_DIR | ENV SENTENCE_TRANSFORMERS_HOME $RAG_EMBEDDING_MODEL_DIR | ||||||
|  |  | ||||||
							
								
								
									
										12
									
								
								README.md
									
										
									
									
									
								
							
							
						
						
									
										12
									
								
								README.md
									
										
									
									
									
								
							|  | @ -53,8 +53,6 @@ User-friendly WebUI for LLMs, supported LLM runners include Ollama and OpenAI-co | ||||||
| 
 | 
 | ||||||
| - 💬 **Collaborative Chat**: Harness the collective intelligence of multiple models by seamlessly orchestrating group conversations. Use the `@` command to specify the model, enabling dynamic and diverse dialogues within your chat interface. Immerse yourself in the collective intelligence woven into your chat environment. | - 💬 **Collaborative Chat**: Harness the collective intelligence of multiple models by seamlessly orchestrating group conversations. Use the `@` command to specify the model, enabling dynamic and diverse dialogues within your chat interface. Immerse yourself in the collective intelligence woven into your chat environment. | ||||||
| 
 | 
 | ||||||
| - 🤝 **OpenAI API Integration**: Effortlessly integrate OpenAI-compatible API for versatile conversations alongside Ollama models. Customize the API Base URL to link with **LMStudio, Mistral, OpenRouter, and more**. |  | ||||||
| 
 |  | ||||||
| - 🔄 **Regeneration History Access**: Easily revisit and explore your entire regeneration history. | - 🔄 **Regeneration History Access**: Easily revisit and explore your entire regeneration history. | ||||||
| 
 | 
 | ||||||
| - 📜 **Chat History**: Effortlessly access and manage your conversation history. | - 📜 **Chat History**: Effortlessly access and manage your conversation history. | ||||||
|  | @ -65,8 +63,18 @@ User-friendly WebUI for LLMs, supported LLM runners include Ollama and OpenAI-co | ||||||
| 
 | 
 | ||||||
| - ⚙️ **Fine-Tuned Control with Advanced Parameters**: Gain a deeper level of control by adjusting parameters such as temperature and defining your system prompts to tailor the conversation to your specific preferences and needs. | - ⚙️ **Fine-Tuned Control with Advanced Parameters**: Gain a deeper level of control by adjusting parameters such as temperature and defining your system prompts to tailor the conversation to your specific preferences and needs. | ||||||
| 
 | 
 | ||||||
|  | - 🎨🤖 **Image Generation Integration**: Seamlessly incorporate image generation capabilities using AUTOMATIC1111 API (local) and DALL-E, enriching your chat experience with dynamic visual content. | ||||||
|  | 
 | ||||||
|  | - 🤝 **OpenAI API Integration**: Effortlessly integrate OpenAI-compatible API for versatile conversations alongside Ollama models. Customize the API Base URL to link with **LMStudio, Mistral, OpenRouter, and more**. | ||||||
|  | 
 | ||||||
|  | - ✨ **Multiple OpenAI-Compatible API Support**: Seamlessly integrate and customize various OpenAI-compatible APIs, enhancing the versatility of your chat interactions. | ||||||
|  | 
 | ||||||
| - 🔗 **External Ollama Server Connection**: Seamlessly link to an external Ollama server hosted on a different address by configuring the environment variable. | - 🔗 **External Ollama Server Connection**: Seamlessly link to an external Ollama server hosted on a different address by configuring the environment variable. | ||||||
| 
 | 
 | ||||||
|  | - 🔀 **Multiple Ollama Instance Load Balancing**: Effortlessly distribute chat requests across multiple Ollama instances for enhanced performance and reliability. | ||||||
|  | 
 | ||||||
|  | - 👥 **Multi-User Management**: Easily oversee and administer users via our intuitive admin panel, streamlining user management processes. | ||||||
|  | 
 | ||||||
| - 🔐 **Role-Based Access Control (RBAC)**: Ensure secure access with restricted permissions; only authorized individuals can access your Ollama, and exclusive model creation/pulling rights are reserved for administrators. | - 🔐 **Role-Based Access Control (RBAC)**: Ensure secure access with restricted permissions; only authorized individuals can access your Ollama, and exclusive model creation/pulling rights are reserved for administrators. | ||||||
| 
 | 
 | ||||||
| - 🔒 **Backend Reverse Proxy Support**: Bolster security through direct communication between Open WebUI backend and Ollama. This key feature eliminates the need to expose Ollama over LAN. Requests made to the '/ollama/api' route from the web UI are seamlessly redirected to Ollama from the backend, enhancing overall system security. | - 🔒 **Backend Reverse Proxy Support**: Bolster security through direct communication between Open WebUI backend and Ollama. This key feature eliminates the need to expose Ollama over LAN. Requests made to the '/ollama/api' route from the web UI are seamlessly redirected to Ollama from the backend, enhancing overall system security. | ||||||
|  |  | ||||||
|  | @ -21,7 +21,16 @@ from utils.utils import ( | ||||||
| from utils.misc import calculate_sha256 | from utils.misc import calculate_sha256 | ||||||
| from typing import Optional | from typing import Optional | ||||||
| from pydantic import BaseModel | from pydantic import BaseModel | ||||||
| from config import AUTOMATIC1111_BASE_URL | from pathlib import Path | ||||||
|  | import uuid | ||||||
|  | import base64 | ||||||
|  | import json | ||||||
|  | 
 | ||||||
|  | from config import CACHE_DIR, AUTOMATIC1111_BASE_URL | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | IMAGE_CACHE_DIR = Path(CACHE_DIR).joinpath("./image/generations/") | ||||||
|  | IMAGE_CACHE_DIR.mkdir(parents=True, exist_ok=True) | ||||||
| 
 | 
 | ||||||
| app = FastAPI() | app = FastAPI() | ||||||
| app.add_middleware( | app.add_middleware( | ||||||
|  | @ -32,25 +41,34 @@ app.add_middleware( | ||||||
|     allow_headers=["*"], |     allow_headers=["*"], | ||||||
| ) | ) | ||||||
| 
 | 
 | ||||||
|  | app.state.ENGINE = "" | ||||||
|  | app.state.ENABLED = False | ||||||
|  | 
 | ||||||
|  | app.state.OPENAI_API_KEY = "" | ||||||
|  | app.state.MODEL = "" | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
| app.state.AUTOMATIC1111_BASE_URL = AUTOMATIC1111_BASE_URL | app.state.AUTOMATIC1111_BASE_URL = AUTOMATIC1111_BASE_URL | ||||||
| app.state.ENABLED = app.state.AUTOMATIC1111_BASE_URL != "" | 
 | ||||||
| app.state.IMAGE_SIZE = "512x512" | app.state.IMAGE_SIZE = "512x512" | ||||||
| app.state.IMAGE_STEPS = 50 | app.state.IMAGE_STEPS = 50 | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| @app.get("/enabled", response_model=bool) | @app.get("/config") | ||||||
| async def get_enable_status(request: Request, user=Depends(get_admin_user)): | async def get_config(request: Request, user=Depends(get_admin_user)): | ||||||
|     return app.state.ENABLED |     return {"engine": app.state.ENGINE, "enabled": app.state.ENABLED} | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| @app.get("/enabled/toggle", response_model=bool) | class ConfigUpdateForm(BaseModel): | ||||||
| async def toggle_enabled(request: Request, user=Depends(get_admin_user)): |     engine: str | ||||||
|     try: |     enabled: bool | ||||||
|         r = requests.head(app.state.AUTOMATIC1111_BASE_URL) | 
 | ||||||
|         app.state.ENABLED = not app.state.ENABLED | 
 | ||||||
|         return app.state.ENABLED | @app.post("/config/update") | ||||||
|     except Exception as e: | async def update_config(form_data: ConfigUpdateForm, user=Depends(get_admin_user)): | ||||||
|         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) |     app.state.ENGINE = form_data.engine | ||||||
|  |     app.state.ENABLED = form_data.enabled | ||||||
|  |     return {"engine": app.state.ENGINE, "enabled": app.state.ENABLED} | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| class UrlUpdateForm(BaseModel): | class UrlUpdateForm(BaseModel): | ||||||
|  | @ -58,17 +76,24 @@ class UrlUpdateForm(BaseModel): | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| @app.get("/url") | @app.get("/url") | ||||||
| async def get_openai_url(user=Depends(get_admin_user)): | async def get_automatic1111_url(user=Depends(get_admin_user)): | ||||||
|     return {"AUTOMATIC1111_BASE_URL": app.state.AUTOMATIC1111_BASE_URL} |     return {"AUTOMATIC1111_BASE_URL": app.state.AUTOMATIC1111_BASE_URL} | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| @app.post("/url/update") | @app.post("/url/update") | ||||||
| async def update_openai_url(form_data: UrlUpdateForm, user=Depends(get_admin_user)): | async def update_automatic1111_url( | ||||||
|  |     form_data: UrlUpdateForm, user=Depends(get_admin_user) | ||||||
|  | ): | ||||||
| 
 | 
 | ||||||
|     if form_data.url == "": |     if form_data.url == "": | ||||||
|         app.state.AUTOMATIC1111_BASE_URL = AUTOMATIC1111_BASE_URL |         app.state.AUTOMATIC1111_BASE_URL = AUTOMATIC1111_BASE_URL | ||||||
|     else: |     else: | ||||||
|         app.state.AUTOMATIC1111_BASE_URL = form_data.url.strip("/") |         url = form_data.url.strip("/") | ||||||
|  |         try: | ||||||
|  |             r = requests.head(url) | ||||||
|  |             app.state.AUTOMATIC1111_BASE_URL = url | ||||||
|  |         except Exception as e: | ||||||
|  |             raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) | ||||||
| 
 | 
 | ||||||
|     return { |     return { | ||||||
|         "AUTOMATIC1111_BASE_URL": app.state.AUTOMATIC1111_BASE_URL, |         "AUTOMATIC1111_BASE_URL": app.state.AUTOMATIC1111_BASE_URL, | ||||||
|  | @ -76,6 +101,30 @@ async def update_openai_url(form_data: UrlUpdateForm, user=Depends(get_admin_use | ||||||
|     } |     } | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | class OpenAIKeyUpdateForm(BaseModel): | ||||||
|  |     key: str | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | @app.get("/key") | ||||||
|  | async def get_openai_key(user=Depends(get_admin_user)): | ||||||
|  |     return {"OPENAI_API_KEY": app.state.OPENAI_API_KEY} | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | @app.post("/key/update") | ||||||
|  | async def update_openai_key( | ||||||
|  |     form_data: OpenAIKeyUpdateForm, user=Depends(get_admin_user) | ||||||
|  | ): | ||||||
|  | 
 | ||||||
|  |     if form_data.key == "": | ||||||
|  |         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.API_KEY_NOT_FOUND) | ||||||
|  | 
 | ||||||
|  |     app.state.OPENAI_API_KEY = form_data.key | ||||||
|  |     return { | ||||||
|  |         "OPENAI_API_KEY": app.state.OPENAI_API_KEY, | ||||||
|  |         "status": True, | ||||||
|  |     } | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
| class ImageSizeUpdateForm(BaseModel): | class ImageSizeUpdateForm(BaseModel): | ||||||
|     size: str |     size: str | ||||||
| 
 | 
 | ||||||
|  | @ -132,9 +181,22 @@ async def update_image_size( | ||||||
| @app.get("/models") | @app.get("/models") | ||||||
| def get_models(user=Depends(get_current_user)): | def get_models(user=Depends(get_current_user)): | ||||||
|     try: |     try: | ||||||
|         r = requests.get(url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/sd-models") |         if app.state.ENGINE == "openai": | ||||||
|         models = r.json() |             return [ | ||||||
|         return models |                 {"id": "dall-e-2", "name": "DALL·E 2"}, | ||||||
|  |                 {"id": "dall-e-3", "name": "DALL·E 3"}, | ||||||
|  |             ] | ||||||
|  |         else: | ||||||
|  |             r = requests.get( | ||||||
|  |                 url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/sd-models" | ||||||
|  |             ) | ||||||
|  |             models = r.json() | ||||||
|  |             return list( | ||||||
|  |                 map( | ||||||
|  |                     lambda model: {"id": model["title"], "name": model["model_name"]}, | ||||||
|  |                     models, | ||||||
|  |                 ) | ||||||
|  |             ) | ||||||
|     except Exception as e: |     except Exception as e: | ||||||
|         app.state.ENABLED = False |         app.state.ENABLED = False | ||||||
|         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) |         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) | ||||||
|  | @ -143,10 +205,12 @@ def get_models(user=Depends(get_current_user)): | ||||||
| @app.get("/models/default") | @app.get("/models/default") | ||||||
| async def get_default_model(user=Depends(get_admin_user)): | async def get_default_model(user=Depends(get_admin_user)): | ||||||
|     try: |     try: | ||||||
|         r = requests.get(url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/options") |         if app.state.ENGINE == "openai": | ||||||
|         options = r.json() |             return {"model": app.state.MODEL if app.state.MODEL else "dall-e-2"} | ||||||
| 
 |         else: | ||||||
|         return {"model": options["sd_model_checkpoint"]} |             r = requests.get(url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/options") | ||||||
|  |             options = r.json() | ||||||
|  |             return {"model": options["sd_model_checkpoint"]} | ||||||
|     except Exception as e: |     except Exception as e: | ||||||
|         app.state.ENABLED = False |         app.state.ENABLED = False | ||||||
|         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) |         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) | ||||||
|  | @ -157,16 +221,21 @@ class UpdateModelForm(BaseModel): | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| def set_model_handler(model: str): | def set_model_handler(model: str): | ||||||
|     r = requests.get(url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/options") |  | ||||||
|     options = r.json() |  | ||||||
| 
 | 
 | ||||||
|     if model != options["sd_model_checkpoint"]: |     if app.state.ENGINE == "openai": | ||||||
|         options["sd_model_checkpoint"] = model |         app.state.MODEL = model | ||||||
|         r = requests.post( |         return app.state.MODEL | ||||||
|             url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/options", json=options |     else: | ||||||
|         ) |         r = requests.get(url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/options") | ||||||
|  |         options = r.json() | ||||||
| 
 | 
 | ||||||
|     return options |         if model != options["sd_model_checkpoint"]: | ||||||
|  |             options["sd_model_checkpoint"] = model | ||||||
|  |             r = requests.post( | ||||||
|  |                 url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/options", json=options | ||||||
|  |             ) | ||||||
|  | 
 | ||||||
|  |         return options | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| @app.post("/models/default/update") | @app.post("/models/default/update") | ||||||
|  | @ -185,6 +254,24 @@ class GenerateImageForm(BaseModel): | ||||||
|     negative_prompt: Optional[str] = None |     negative_prompt: Optional[str] = None | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | def save_b64_image(b64_str): | ||||||
|  |     image_id = str(uuid.uuid4()) | ||||||
|  |     file_path = IMAGE_CACHE_DIR.joinpath(f"{image_id}.png") | ||||||
|  | 
 | ||||||
|  |     try: | ||||||
|  |         # Split the base64 string to get the actual image data | ||||||
|  |         img_data = base64.b64decode(b64_str) | ||||||
|  | 
 | ||||||
|  |         # Write the image data to a file | ||||||
|  |         with open(file_path, "wb") as f: | ||||||
|  |             f.write(img_data) | ||||||
|  | 
 | ||||||
|  |         return image_id | ||||||
|  |     except Exception as e: | ||||||
|  |         print(f"Error saving image: {e}") | ||||||
|  |         return None | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
| @app.post("/generations") | @app.post("/generations") | ||||||
| def generate_image( | def generate_image( | ||||||
|     form_data: GenerateImageForm, |     form_data: GenerateImageForm, | ||||||
|  | @ -194,32 +281,82 @@ def generate_image( | ||||||
|     print(form_data) |     print(form_data) | ||||||
| 
 | 
 | ||||||
|     try: |     try: | ||||||
|         if form_data.model: |         if app.state.ENGINE == "openai": | ||||||
|             set_model_handler(form_data.model) |  | ||||||
| 
 | 
 | ||||||
|         width, height = tuple(map(int, app.state.IMAGE_SIZE.split("x"))) |             headers = {} | ||||||
|  |             headers["Authorization"] = f"Bearer {app.state.OPENAI_API_KEY}" | ||||||
|  |             headers["Content-Type"] = "application/json" | ||||||
| 
 | 
 | ||||||
|         data = { |             data = { | ||||||
|             "prompt": form_data.prompt, |                 "model": app.state.MODEL if app.state.MODEL != "" else "dall-e-2", | ||||||
|             "batch_size": form_data.n, |                 "prompt": form_data.prompt, | ||||||
|             "width": width, |                 "n": form_data.n, | ||||||
|             "height": height, |                 "size": form_data.size, | ||||||
|         } |                 "response_format": "b64_json", | ||||||
|  |             } | ||||||
| 
 | 
 | ||||||
|         if app.state.IMAGE_STEPS != None: |             r = requests.post( | ||||||
|             data["steps"] = app.state.IMAGE_STEPS |                 url=f"https://api.openai.com/v1/images/generations", | ||||||
|  |                 json=data, | ||||||
|  |                 headers=headers, | ||||||
|  |             ) | ||||||
| 
 | 
 | ||||||
|         if form_data.negative_prompt != None: |             r.raise_for_status() | ||||||
|             data["negative_prompt"] = form_data.negative_prompt |  | ||||||
| 
 | 
 | ||||||
|         print(data) |             res = r.json() | ||||||
| 
 | 
 | ||||||
|         r = requests.post( |             images = [] | ||||||
|             url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/txt2img", | 
 | ||||||
|             json=data, |             for image in res["data"]: | ||||||
|         ) |                 image_id = save_b64_image(image["b64_json"]) | ||||||
|  |                 images.append({"url": f"/cache/image/generations/{image_id}.png"}) | ||||||
|  |                 file_body_path = IMAGE_CACHE_DIR.joinpath(f"{image_id}.json") | ||||||
|  | 
 | ||||||
|  |                 with open(file_body_path, "w") as f: | ||||||
|  |                     json.dump(data, f) | ||||||
|  | 
 | ||||||
|  |             return images | ||||||
|  | 
 | ||||||
|  |         else: | ||||||
|  |             if form_data.model: | ||||||
|  |                 set_model_handler(form_data.model) | ||||||
|  | 
 | ||||||
|  |             width, height = tuple(map(int, app.state.IMAGE_SIZE.split("x"))) | ||||||
|  | 
 | ||||||
|  |             data = { | ||||||
|  |                 "prompt": form_data.prompt, | ||||||
|  |                 "batch_size": form_data.n, | ||||||
|  |                 "width": width, | ||||||
|  |                 "height": height, | ||||||
|  |             } | ||||||
|  | 
 | ||||||
|  |             if app.state.IMAGE_STEPS != None: | ||||||
|  |                 data["steps"] = app.state.IMAGE_STEPS | ||||||
|  | 
 | ||||||
|  |             if form_data.negative_prompt != None: | ||||||
|  |                 data["negative_prompt"] = form_data.negative_prompt | ||||||
|  | 
 | ||||||
|  |             r = requests.post( | ||||||
|  |                 url=f"{app.state.AUTOMATIC1111_BASE_URL}/sdapi/v1/txt2img", | ||||||
|  |                 json=data, | ||||||
|  |             ) | ||||||
|  | 
 | ||||||
|  |             res = r.json() | ||||||
|  | 
 | ||||||
|  |             print(res) | ||||||
|  | 
 | ||||||
|  |             images = [] | ||||||
|  | 
 | ||||||
|  |             for image in res["images"]: | ||||||
|  |                 image_id = save_b64_image(image) | ||||||
|  |                 images.append({"url": f"/cache/image/generations/{image_id}.png"}) | ||||||
|  |                 file_body_path = IMAGE_CACHE_DIR.joinpath(f"{image_id}.json") | ||||||
|  | 
 | ||||||
|  |                 with open(file_body_path, "w") as f: | ||||||
|  |                     json.dump({**data, "info": res["info"]}, f) | ||||||
|  | 
 | ||||||
|  |             return images | ||||||
| 
 | 
 | ||||||
|         return r.json() |  | ||||||
|     except Exception as e: |     except Exception as e: | ||||||
|         print(e) |         print(e) | ||||||
|         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) |         raise HTTPException(status_code=400, detail=ERROR_MESSAGES.DEFAULT(e)) | ||||||
|  |  | ||||||
							
								
								
									
										41
									
								
								backend/apps/litellm/main.py
									
										
									
									
									
										Normal file
									
								
							
							
						
						
									
										41
									
								
								backend/apps/litellm/main.py
									
										
									
									
									
										Normal file
									
								
							|  | @ -0,0 +1,41 @@ | ||||||
|  | from litellm.proxy.proxy_server import ProxyConfig, initialize | ||||||
|  | from litellm.proxy.proxy_server import app | ||||||
|  | 
 | ||||||
|  | from fastapi import FastAPI, Request, Depends, status | ||||||
|  | from fastapi.responses import JSONResponse | ||||||
|  | from utils.utils import get_http_authorization_cred, get_current_user | ||||||
|  | from config import ENV | ||||||
|  | 
 | ||||||
|  | proxy_config = ProxyConfig() | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | async def config(): | ||||||
|  |     router, model_list, general_settings = await proxy_config.load_config( | ||||||
|  |         router=None, config_file_path="./data/litellm/config.yaml" | ||||||
|  |     ) | ||||||
|  | 
 | ||||||
|  |     await initialize(config="./data/litellm/config.yaml", telemetry=False) | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | async def startup(): | ||||||
|  |     await config() | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | @app.on_event("startup") | ||||||
|  | async def on_startup(): | ||||||
|  |     await startup() | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | @app.middleware("http") | ||||||
|  | async def auth_middleware(request: Request, call_next): | ||||||
|  |     auth_header = request.headers.get("Authorization", "") | ||||||
|  | 
 | ||||||
|  |     if ENV != "dev": | ||||||
|  |         try: | ||||||
|  |             user = get_current_user(get_http_authorization_cred(auth_header)) | ||||||
|  |             print(user) | ||||||
|  |         except Exception as e: | ||||||
|  |             return JSONResponse(status_code=400, content={"detail": str(e)}) | ||||||
|  | 
 | ||||||
|  |     response = await call_next(request) | ||||||
|  |     return response | ||||||
|  | @ -9,17 +9,14 @@ import requests | ||||||
| from fastapi import FastAPI, Request, Depends, status | from fastapi import FastAPI, Request, Depends, status | ||||||
| from fastapi.staticfiles import StaticFiles | from fastapi.staticfiles import StaticFiles | ||||||
| from fastapi import HTTPException | from fastapi import HTTPException | ||||||
| from fastapi.responses import JSONResponse |  | ||||||
| from fastapi.middleware.wsgi import WSGIMiddleware | from fastapi.middleware.wsgi import WSGIMiddleware | ||||||
| from fastapi.middleware.cors import CORSMiddleware | from fastapi.middleware.cors import CORSMiddleware | ||||||
| from starlette.exceptions import HTTPException as StarletteHTTPException | from starlette.exceptions import HTTPException as StarletteHTTPException | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| from litellm.proxy.proxy_server import ProxyConfig, initialize |  | ||||||
| from litellm.proxy.proxy_server import app as litellm_app |  | ||||||
| 
 |  | ||||||
| from apps.ollama.main import app as ollama_app | from apps.ollama.main import app as ollama_app | ||||||
| from apps.openai.main import app as openai_app | from apps.openai.main import app as openai_app | ||||||
|  | from apps.litellm.main import app as litellm_app, startup as litellm_app_startup | ||||||
| from apps.audio.main import app as audio_app | from apps.audio.main import app as audio_app | ||||||
| from apps.images.main import app as images_app | from apps.images.main import app as images_app | ||||||
| from apps.rag.main import app as rag_app | from apps.rag.main import app as rag_app | ||||||
|  | @ -29,8 +26,6 @@ from apps.web.main import app as webui_app | ||||||
| from config import WEBUI_NAME, ENV, VERSION, CHANGELOG, FRONTEND_BUILD_DIR | from config import WEBUI_NAME, ENV, VERSION, CHANGELOG, FRONTEND_BUILD_DIR | ||||||
| from constants import ERROR_MESSAGES | from constants import ERROR_MESSAGES | ||||||
| 
 | 
 | ||||||
| from utils.utils import get_http_authorization_cred, get_current_user |  | ||||||
| 
 |  | ||||||
| 
 | 
 | ||||||
| class SPAStaticFiles(StaticFiles): | class SPAStaticFiles(StaticFiles): | ||||||
|     async def get_response(self, path: str, scope): |     async def get_response(self, path: str, scope): | ||||||
|  | @ -43,21 +38,6 @@ class SPAStaticFiles(StaticFiles): | ||||||
|                 raise ex |                 raise ex | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| proxy_config = ProxyConfig() |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| async def config(): |  | ||||||
|     router, model_list, general_settings = await proxy_config.load_config( |  | ||||||
|         router=None, config_file_path="./data/litellm/config.yaml" |  | ||||||
|     ) |  | ||||||
| 
 |  | ||||||
|     await initialize(config="./data/litellm/config.yaml", telemetry=False) |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| async def startup(): |  | ||||||
|     await config() |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| app = FastAPI(docs_url="/docs" if ENV == "dev" else None, redoc_url=None) | app = FastAPI(docs_url="/docs" if ENV == "dev" else None, redoc_url=None) | ||||||
| 
 | 
 | ||||||
| origins = ["*"] | origins = ["*"] | ||||||
|  | @ -73,7 +53,7 @@ app.add_middleware( | ||||||
| 
 | 
 | ||||||
| @app.on_event("startup") | @app.on_event("startup") | ||||||
| async def on_startup(): | async def on_startup(): | ||||||
|     await startup() |     await litellm_app_startup() | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| @app.middleware("http") | @app.middleware("http") | ||||||
|  | @ -86,21 +66,6 @@ async def check_url(request: Request, call_next): | ||||||
|     return response |     return response | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| @litellm_app.middleware("http") |  | ||||||
| async def auth_middleware(request: Request, call_next): |  | ||||||
|     auth_header = request.headers.get("Authorization", "") |  | ||||||
| 
 |  | ||||||
|     if ENV != "dev": |  | ||||||
|         try: |  | ||||||
|             user = get_current_user(get_http_authorization_cred(auth_header)) |  | ||||||
|             print(user) |  | ||||||
|         except Exception as e: |  | ||||||
|             return JSONResponse(status_code=400, content={"detail": str(e)}) |  | ||||||
| 
 |  | ||||||
|     response = await call_next(request) |  | ||||||
|     return response |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| app.mount("/api/v1", webui_app) | app.mount("/api/v1", webui_app) | ||||||
| app.mount("/litellm/api", litellm_app) | app.mount("/litellm/api", litellm_app) | ||||||
| 
 | 
 | ||||||
|  | @ -156,6 +121,7 @@ async def get_app_latest_release_version(): | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| app.mount("/static", StaticFiles(directory="static"), name="static") | app.mount("/static", StaticFiles(directory="static"), name="static") | ||||||
|  | app.mount("/cache", StaticFiles(directory="data/cache"), name="cache") | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| app.mount( | app.mount( | ||||||
|  |  | ||||||
|  | @ -1,9 +1,9 @@ | ||||||
| import { IMAGES_API_BASE_URL } from '$lib/constants'; | import { IMAGES_API_BASE_URL } from '$lib/constants'; | ||||||
| 
 | 
 | ||||||
| export const getImageGenerationEnabledStatus = async (token: string = '') => { | export const getImageGenerationConfig = async (token: string = '') => { | ||||||
| 	let error = null; | 	let error = null; | ||||||
| 
 | 
 | ||||||
| 	const res = await fetch(`${IMAGES_API_BASE_URL}/enabled`, { | 	const res = await fetch(`${IMAGES_API_BASE_URL}/config`, { | ||||||
| 		method: 'GET', | 		method: 'GET', | ||||||
| 		headers: { | 		headers: { | ||||||
| 			Accept: 'application/json', | 			Accept: 'application/json', | ||||||
|  | @ -32,10 +32,50 @@ export const getImageGenerationEnabledStatus = async (token: string = '') => { | ||||||
| 	return res; | 	return res; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| export const toggleImageGenerationEnabledStatus = async (token: string = '') => { | export const updateImageGenerationConfig = async ( | ||||||
|  | 	token: string = '', | ||||||
|  | 	engine: string, | ||||||
|  | 	enabled: boolean | ||||||
|  | ) => { | ||||||
| 	let error = null; | 	let error = null; | ||||||
| 
 | 
 | ||||||
| 	const res = await fetch(`${IMAGES_API_BASE_URL}/enabled/toggle`, { | 	const res = await fetch(`${IMAGES_API_BASE_URL}/config/update`, { | ||||||
|  | 		method: 'POST', | ||||||
|  | 		headers: { | ||||||
|  | 			Accept: 'application/json', | ||||||
|  | 			'Content-Type': 'application/json', | ||||||
|  | 			...(token && { authorization: `Bearer ${token}` }) | ||||||
|  | 		}, | ||||||
|  | 		body: JSON.stringify({ | ||||||
|  | 			engine, | ||||||
|  | 			enabled | ||||||
|  | 		}) | ||||||
|  | 	}) | ||||||
|  | 		.then(async (res) => { | ||||||
|  | 			if (!res.ok) throw await res.json(); | ||||||
|  | 			return res.json(); | ||||||
|  | 		}) | ||||||
|  | 		.catch((err) => { | ||||||
|  | 			console.log(err); | ||||||
|  | 			if ('detail' in err) { | ||||||
|  | 				error = err.detail; | ||||||
|  | 			} else { | ||||||
|  | 				error = 'Server connection failed'; | ||||||
|  | 			} | ||||||
|  | 			return null; | ||||||
|  | 		}); | ||||||
|  | 
 | ||||||
|  | 	if (error) { | ||||||
|  | 		throw error; | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return res; | ||||||
|  | }; | ||||||
|  | 
 | ||||||
|  | export const getOpenAIKey = async (token: string = '') => { | ||||||
|  | 	let error = null; | ||||||
|  | 
 | ||||||
|  | 	const res = await fetch(`${IMAGES_API_BASE_URL}/key`, { | ||||||
| 		method: 'GET', | 		method: 'GET', | ||||||
| 		headers: { | 		headers: { | ||||||
| 			Accept: 'application/json', | 			Accept: 'application/json', | ||||||
|  | @ -61,7 +101,42 @@ export const toggleImageGenerationEnabledStatus = async (token: string = '') => | ||||||
| 		throw error; | 		throw error; | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	return res; | 	return res.OPENAI_API_KEY; | ||||||
|  | }; | ||||||
|  | 
 | ||||||
|  | export const updateOpenAIKey = async (token: string = '', key: string) => { | ||||||
|  | 	let error = null; | ||||||
|  | 
 | ||||||
|  | 	const res = await fetch(`${IMAGES_API_BASE_URL}/key/update`, { | ||||||
|  | 		method: 'POST', | ||||||
|  | 		headers: { | ||||||
|  | 			Accept: 'application/json', | ||||||
|  | 			'Content-Type': 'application/json', | ||||||
|  | 			...(token && { authorization: `Bearer ${token}` }) | ||||||
|  | 		}, | ||||||
|  | 		body: JSON.stringify({ | ||||||
|  | 			key: key | ||||||
|  | 		}) | ||||||
|  | 	}) | ||||||
|  | 		.then(async (res) => { | ||||||
|  | 			if (!res.ok) throw await res.json(); | ||||||
|  | 			return res.json(); | ||||||
|  | 		}) | ||||||
|  | 		.catch((err) => { | ||||||
|  | 			console.log(err); | ||||||
|  | 			if ('detail' in err) { | ||||||
|  | 				error = err.detail; | ||||||
|  | 			} else { | ||||||
|  | 				error = 'Server connection failed'; | ||||||
|  | 			} | ||||||
|  | 			return null; | ||||||
|  | 		}); | ||||||
|  | 
 | ||||||
|  | 	if (error) { | ||||||
|  | 		throw error; | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return res.OPENAI_API_KEY; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| export const getAUTOMATIC1111Url = async (token: string = '') => { | export const getAUTOMATIC1111Url = async (token: string = '') => { | ||||||
|  | @ -263,7 +338,7 @@ export const updateImageSteps = async (token: string = '', steps: number) => { | ||||||
| 	return res.IMAGE_STEPS; | 	return res.IMAGE_STEPS; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| export const getDiffusionModels = async (token: string = '') => { | export const getImageGenerationModels = async (token: string = '') => { | ||||||
| 	let error = null; | 	let error = null; | ||||||
| 
 | 
 | ||||||
| 	const res = await fetch(`${IMAGES_API_BASE_URL}/models`, { | 	const res = await fetch(`${IMAGES_API_BASE_URL}/models`, { | ||||||
|  | @ -295,7 +370,7 @@ export const getDiffusionModels = async (token: string = '') => { | ||||||
| 	return res; | 	return res; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| export const getDefaultDiffusionModel = async (token: string = '') => { | export const getDefaultImageGenerationModel = async (token: string = '') => { | ||||||
| 	let error = null; | 	let error = null; | ||||||
| 
 | 
 | ||||||
| 	const res = await fetch(`${IMAGES_API_BASE_URL}/models/default`, { | 	const res = await fetch(`${IMAGES_API_BASE_URL}/models/default`, { | ||||||
|  | @ -327,7 +402,7 @@ export const getDefaultDiffusionModel = async (token: string = '') => { | ||||||
| 	return res.model; | 	return res.model; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| export const updateDefaultDiffusionModel = async (token: string = '', model: string) => { | export const updateDefaultImageGenerationModel = async (token: string = '', model: string) => { | ||||||
| 	let error = null; | 	let error = null; | ||||||
| 
 | 
 | ||||||
| 	const res = await fetch(`${IMAGES_API_BASE_URL}/models/default/update`, { | 	const res = await fetch(`${IMAGES_API_BASE_URL}/models/default/update`, { | ||||||
|  |  | ||||||
|  | @ -77,6 +77,7 @@ type AddLiteLLMModelForm = { | ||||||
| 	api_base: string; | 	api_base: string; | ||||||
| 	api_key: string; | 	api_key: string; | ||||||
| 	rpm: string; | 	rpm: string; | ||||||
|  | 	max_tokens: string; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| export const addLiteLLMModel = async (token: string = '', payload: AddLiteLLMModelForm) => { | export const addLiteLLMModel = async (token: string = '', payload: AddLiteLLMModelForm) => { | ||||||
|  | @ -95,7 +96,8 @@ export const addLiteLLMModel = async (token: string = '', payload: AddLiteLLMMod | ||||||
| 				model: payload.model, | 				model: payload.model, | ||||||
| 				...(payload.api_base === '' ? {} : { api_base: payload.api_base }), | 				...(payload.api_base === '' ? {} : { api_base: payload.api_base }), | ||||||
| 				...(payload.api_key === '' ? {} : { api_key: payload.api_key }), | 				...(payload.api_key === '' ? {} : { api_key: payload.api_key }), | ||||||
| 				...(isNaN(parseInt(payload.rpm)) ? {} : { rpm: parseInt(payload.rpm) }) | 				...(isNaN(parseInt(payload.rpm)) ? {} : { rpm: parseInt(payload.rpm) }), | ||||||
|  | 				...(payload.max_tokens === '' ? {} : { max_tokens: payload.max_tokens }) | ||||||
| 			} | 			} | ||||||
| 		}) | 		}) | ||||||
| 	}) | 	}) | ||||||
|  |  | ||||||
|  | @ -6,6 +6,7 @@ | ||||||
| 	import auto_render from 'katex/dist/contrib/auto-render.mjs'; | 	import auto_render from 'katex/dist/contrib/auto-render.mjs'; | ||||||
| 	import 'katex/dist/katex.min.css'; | 	import 'katex/dist/katex.min.css'; | ||||||
| 
 | 
 | ||||||
|  | 	import { fade } from 'svelte/transition'; | ||||||
| 	import { createEventDispatcher } from 'svelte'; | 	import { createEventDispatcher } from 'svelte'; | ||||||
| 	import { onMount, tick, getContext } from 'svelte'; | 	import { onMount, tick, getContext } from 'svelte'; | ||||||
| 
 | 
 | ||||||
|  | @ -278,13 +279,15 @@ | ||||||
| 
 | 
 | ||||||
| 	const generateImage = async (message) => { | 	const generateImage = async (message) => { | ||||||
| 		generatingImage = true; | 		generatingImage = true; | ||||||
| 		const res = await imageGenerations(localStorage.token, message.content); | 		const res = await imageGenerations(localStorage.token, message.content).catch((error) => { | ||||||
|  | 			toast.error(error); | ||||||
|  | 		}); | ||||||
| 		console.log(res); | 		console.log(res); | ||||||
| 
 | 
 | ||||||
| 		if (res) { | 		if (res) { | ||||||
| 			message.files = res.images.map((image) => ({ | 			message.files = res.map((image) => ({ | ||||||
| 				type: 'image', | 				type: 'image', | ||||||
| 				url: `data:image/png;base64,${image}` | 				url: `${image.url}` | ||||||
| 			})); | 			})); | ||||||
| 
 | 
 | ||||||
| 			dispatch('save', message); | 			dispatch('save', message); | ||||||
|  | @ -479,7 +482,7 @@ | ||||||
| 													xmlns="http://www.w3.org/2000/svg" | 													xmlns="http://www.w3.org/2000/svg" | ||||||
| 													fill="none" | 													fill="none" | ||||||
| 													viewBox="0 0 24 24" | 													viewBox="0 0 24 24" | ||||||
| 													stroke-width="1.5" | 													stroke-width="2" | ||||||
| 													stroke="currentColor" | 													stroke="currentColor" | ||||||
| 													class="w-4 h-4" | 													class="w-4 h-4" | ||||||
| 												> | 												> | ||||||
|  | @ -505,7 +508,7 @@ | ||||||
| 													xmlns="http://www.w3.org/2000/svg" | 													xmlns="http://www.w3.org/2000/svg" | ||||||
| 													fill="none" | 													fill="none" | ||||||
| 													viewBox="0 0 24 24" | 													viewBox="0 0 24 24" | ||||||
| 													stroke-width="1.5" | 													stroke-width="2" | ||||||
| 													stroke="currentColor" | 													stroke="currentColor" | ||||||
| 													class="w-4 h-4" | 													class="w-4 h-4" | ||||||
| 												> | 												> | ||||||
|  | @ -624,7 +627,7 @@ | ||||||
| 														xmlns="http://www.w3.org/2000/svg" | 														xmlns="http://www.w3.org/2000/svg" | ||||||
| 														fill="none" | 														fill="none" | ||||||
| 														viewBox="0 0 24 24" | 														viewBox="0 0 24 24" | ||||||
| 														stroke-width="1.5" | 														stroke-width="2" | ||||||
| 														stroke="currentColor" | 														stroke="currentColor" | ||||||
| 														class="w-4 h-4" | 														class="w-4 h-4" | ||||||
| 													> | 													> | ||||||
|  | @ -639,7 +642,7 @@ | ||||||
| 														xmlns="http://www.w3.org/2000/svg" | 														xmlns="http://www.w3.org/2000/svg" | ||||||
| 														fill="none" | 														fill="none" | ||||||
| 														viewBox="0 0 24 24" | 														viewBox="0 0 24 24" | ||||||
| 														stroke-width="1.5" | 														stroke-width="2" | ||||||
| 														stroke="currentColor" | 														stroke="currentColor" | ||||||
| 														class="w-4 h-4" | 														class="w-4 h-4" | ||||||
| 													> | 													> | ||||||
|  | @ -705,7 +708,7 @@ | ||||||
| 															xmlns="http://www.w3.org/2000/svg" | 															xmlns="http://www.w3.org/2000/svg" | ||||||
| 															fill="none" | 															fill="none" | ||||||
| 															viewBox="0 0 24 24" | 															viewBox="0 0 24 24" | ||||||
| 															stroke-width="1.5" | 															stroke-width="2" | ||||||
| 															stroke="currentColor" | 															stroke="currentColor" | ||||||
| 															class="w-4 h-4" | 															class="w-4 h-4" | ||||||
| 														> | 														> | ||||||
|  | @ -735,7 +738,7 @@ | ||||||
| 														xmlns="http://www.w3.org/2000/svg" | 														xmlns="http://www.w3.org/2000/svg" | ||||||
| 														fill="none" | 														fill="none" | ||||||
| 														viewBox="0 0 24 24" | 														viewBox="0 0 24 24" | ||||||
| 														stroke-width="1.5" | 														stroke-width="2" | ||||||
| 														stroke="currentColor" | 														stroke="currentColor" | ||||||
| 														class="w-4 h-4" | 														class="w-4 h-4" | ||||||
| 													> | 													> | ||||||
|  | @ -764,7 +767,7 @@ | ||||||
| 														xmlns="http://www.w3.org/2000/svg" | 														xmlns="http://www.w3.org/2000/svg" | ||||||
| 														fill="none" | 														fill="none" | ||||||
| 														viewBox="0 0 24 24" | 														viewBox="0 0 24 24" | ||||||
| 														stroke-width="1.5" | 														stroke-width="2" | ||||||
| 														stroke="currentColor" | 														stroke="currentColor" | ||||||
| 														class="w-4 h-4" | 														class="w-4 h-4" | ||||||
| 													> | 													> | ||||||
|  | @ -794,7 +797,7 @@ | ||||||
| 														xmlns="http://www.w3.org/2000/svg" | 														xmlns="http://www.w3.org/2000/svg" | ||||||
| 														fill="none" | 														fill="none" | ||||||
| 														viewBox="0 0 24 24" | 														viewBox="0 0 24 24" | ||||||
| 														stroke-width="1.5" | 														stroke-width="2" | ||||||
| 														stroke="currentColor" | 														stroke="currentColor" | ||||||
| 														class="w-4 h-4" | 														class="w-4 h-4" | ||||||
| 													> | 													> | ||||||
|  |  | ||||||
|  | @ -261,7 +261,7 @@ | ||||||
| 									xmlns="http://www.w3.org/2000/svg" | 									xmlns="http://www.w3.org/2000/svg" | ||||||
| 									fill="none" | 									fill="none" | ||||||
| 									viewBox="0 0 24 24" | 									viewBox="0 0 24 24" | ||||||
| 									stroke-width="1.5" | 									stroke-width="2" | ||||||
| 									stroke="currentColor" | 									stroke="currentColor" | ||||||
| 									class="w-4 h-4" | 									class="w-4 h-4" | ||||||
| 								> | 								> | ||||||
|  | @ -285,7 +285,7 @@ | ||||||
| 									xmlns="http://www.w3.org/2000/svg" | 									xmlns="http://www.w3.org/2000/svg" | ||||||
| 									fill="none" | 									fill="none" | ||||||
| 									viewBox="0 0 24 24" | 									viewBox="0 0 24 24" | ||||||
| 									stroke-width="1.5" | 									stroke-width="2" | ||||||
| 									stroke="currentColor" | 									stroke="currentColor" | ||||||
| 									class="w-4 h-4" | 									class="w-4 h-4" | ||||||
| 								> | 								> | ||||||
|  | @ -310,7 +310,7 @@ | ||||||
| 										xmlns="http://www.w3.org/2000/svg" | 										xmlns="http://www.w3.org/2000/svg" | ||||||
| 										fill="none" | 										fill="none" | ||||||
| 										viewBox="0 0 24 24" | 										viewBox="0 0 24 24" | ||||||
| 										stroke-width="1.5" | 										stroke-width="2" | ||||||
| 										stroke="currentColor" | 										stroke="currentColor" | ||||||
| 										class="w-4 h-4" | 										class="w-4 h-4" | ||||||
| 									> | 									> | ||||||
|  |  | ||||||
|  | @ -186,7 +186,7 @@ | ||||||
| 			<div class=" my-2.5 text-sm font-medium">{$i18n.t('System Prompt')}</div> | 			<div class=" my-2.5 text-sm font-medium">{$i18n.t('System Prompt')}</div> | ||||||
| 			<textarea | 			<textarea | ||||||
| 				bind:value={system} | 				bind:value={system} | ||||||
| 				class="w-full rounded p-4 text-sm dark:text-gray-300 dark:bg-gray-800 outline-none resize-none" | 				class="w-full rounded-lg p-4 text-sm dark:text-gray-300 dark:bg-gray-850 outline-none resize-none" | ||||||
| 				rows="4" | 				rows="4" | ||||||
| 			/> | 			/> | ||||||
| 		</div> | 		</div> | ||||||
|  |  | ||||||
|  | @ -5,16 +5,18 @@ | ||||||
| 	import { config, user } from '$lib/stores'; | 	import { config, user } from '$lib/stores'; | ||||||
| 	import { | 	import { | ||||||
| 		getAUTOMATIC1111Url, | 		getAUTOMATIC1111Url, | ||||||
| 		getDefaultDiffusionModel, | 		getImageGenerationModels, | ||||||
| 		getDiffusionModels, | 		getDefaultImageGenerationModel, | ||||||
| 		getImageGenerationEnabledStatus, | 		updateDefaultImageGenerationModel, | ||||||
| 		getImageSize, | 		getImageSize, | ||||||
| 		toggleImageGenerationEnabledStatus, | 		getImageGenerationConfig, | ||||||
|  | 		updateImageGenerationConfig, | ||||||
| 		updateAUTOMATIC1111Url, | 		updateAUTOMATIC1111Url, | ||||||
| 		updateDefaultDiffusionModel, |  | ||||||
| 		updateImageSize, | 		updateImageSize, | ||||||
| 		getImageSteps, | 		getImageSteps, | ||||||
| 		updateImageSteps | 		updateImageSteps, | ||||||
|  | 		getOpenAIKey, | ||||||
|  | 		updateOpenAIKey | ||||||
| 	} from '$lib/apis/images'; | 	} from '$lib/apis/images'; | ||||||
| 	import { getBackendConfig } from '$lib/apis'; | 	import { getBackendConfig } from '$lib/apis'; | ||||||
| 	const dispatch = createEventDispatcher(); | 	const dispatch = createEventDispatcher(); | ||||||
|  | @ -25,8 +27,11 @@ | ||||||
| 
 | 
 | ||||||
| 	let loading = false; | 	let loading = false; | ||||||
| 
 | 
 | ||||||
|  | 	let imageGenerationEngine = ''; | ||||||
| 	let enableImageGeneration = false; | 	let enableImageGeneration = false; | ||||||
|  | 
 | ||||||
| 	let AUTOMATIC1111_BASE_URL = ''; | 	let AUTOMATIC1111_BASE_URL = ''; | ||||||
|  | 	let OPENAI_API_KEY = ''; | ||||||
| 
 | 
 | ||||||
| 	let selectedModel = ''; | 	let selectedModel = ''; | ||||||
| 	let models = null; | 	let models = null; | ||||||
|  | @ -35,11 +40,11 @@ | ||||||
| 	let steps = 50; | 	let steps = 50; | ||||||
| 
 | 
 | ||||||
| 	const getModels = async () => { | 	const getModels = async () => { | ||||||
| 		models = await getDiffusionModels(localStorage.token).catch((error) => { | 		models = await getImageGenerationModels(localStorage.token).catch((error) => { | ||||||
| 			toast.error(error); | 			toast.error(error); | ||||||
| 			return null; | 			return null; | ||||||
| 		}); | 		}); | ||||||
| 		selectedModel = await getDefaultDiffusionModel(localStorage.token).catch((error) => { | 		selectedModel = await getDefaultImageGenerationModel(localStorage.token).catch((error) => { | ||||||
| 			return ''; | 			return ''; | ||||||
| 		}); | 		}); | ||||||
| 	}; | 	}; | ||||||
|  | @ -64,33 +69,45 @@ | ||||||
| 			AUTOMATIC1111_BASE_URL = await getAUTOMATIC1111Url(localStorage.token); | 			AUTOMATIC1111_BASE_URL = await getAUTOMATIC1111Url(localStorage.token); | ||||||
| 		} | 		} | ||||||
| 	}; | 	}; | ||||||
| 	const toggleImageGeneration = async () => { | 	const updateImageGeneration = async () => { | ||||||
| 		if (AUTOMATIC1111_BASE_URL) { | 		const res = await updateImageGenerationConfig( | ||||||
| 			enableImageGeneration = await toggleImageGenerationEnabledStatus(localStorage.token).catch( | 			localStorage.token, | ||||||
| 				(error) => { | 			imageGenerationEngine, | ||||||
| 					toast.error(error); | 			enableImageGeneration | ||||||
| 					return false; | 		).catch((error) => { | ||||||
| 				} | 			toast.error(error); | ||||||
| 			); | 			return null; | ||||||
|  | 		}); | ||||||
| 
 | 
 | ||||||
| 			if (enableImageGeneration) { | 		if (res) { | ||||||
| 				config.set(await getBackendConfig(localStorage.token)); | 			imageGenerationEngine = res.engine; | ||||||
| 				getModels(); | 			enableImageGeneration = res.enabled; | ||||||
| 			} | 		} | ||||||
| 		} else { | 
 | ||||||
| 			enableImageGeneration = false; | 		if (enableImageGeneration) { | ||||||
| 			toast.error($i18n.t('{{item}} not provided', { item: 'AUTOMATIC1111_BASE_URL' })); | 			config.set(await getBackendConfig(localStorage.token)); | ||||||
|  | 			getModels(); | ||||||
| 		} | 		} | ||||||
| 	}; | 	}; | ||||||
| 
 | 
 | ||||||
| 	onMount(async () => { | 	onMount(async () => { | ||||||
| 		if ($user.role === 'admin') { | 		if ($user.role === 'admin') { | ||||||
| 			enableImageGeneration = await getImageGenerationEnabledStatus(localStorage.token); | 			const res = await getImageGenerationConfig(localStorage.token).catch((error) => { | ||||||
| 			AUTOMATIC1111_BASE_URL = await getAUTOMATIC1111Url(localStorage.token); | 				toast.error(error); | ||||||
|  | 				return null; | ||||||
|  | 			}); | ||||||
| 
 | 
 | ||||||
| 			if (enableImageGeneration && AUTOMATIC1111_BASE_URL) { | 			if (res) { | ||||||
| 				imageSize = await getImageSize(localStorage.token); | 				imageGenerationEngine = res.engine; | ||||||
| 				steps = await getImageSteps(localStorage.token); | 				enableImageGeneration = res.enabled; | ||||||
|  | 			} | ||||||
|  | 			AUTOMATIC1111_BASE_URL = await getAUTOMATIC1111Url(localStorage.token); | ||||||
|  | 			OPENAI_API_KEY = await getOpenAIKey(localStorage.token); | ||||||
|  | 
 | ||||||
|  | 			imageSize = await getImageSize(localStorage.token); | ||||||
|  | 			steps = await getImageSteps(localStorage.token); | ||||||
|  | 
 | ||||||
|  | 			if (enableImageGeneration) { | ||||||
| 				getModels(); | 				getModels(); | ||||||
| 			} | 			} | ||||||
| 		} | 		} | ||||||
|  | @ -101,7 +118,11 @@ | ||||||
| 	class="flex flex-col h-full justify-between space-y-3 text-sm" | 	class="flex flex-col h-full justify-between space-y-3 text-sm" | ||||||
| 	on:submit|preventDefault={async () => { | 	on:submit|preventDefault={async () => { | ||||||
| 		loading = true; | 		loading = true; | ||||||
| 		await updateDefaultDiffusionModel(localStorage.token, selectedModel); | 		await updateOpenAIKey(localStorage.token, OPENAI_API_KEY); | ||||||
|  | 
 | ||||||
|  | 		await updateDefaultImageGenerationModel(localStorage.token, selectedModel); | ||||||
|  | 
 | ||||||
|  | 		await updateDefaultImageGenerationModel(localStorage.token, selectedModel); | ||||||
| 		await updateImageSize(localStorage.token, imageSize).catch((error) => { | 		await updateImageSize(localStorage.token, imageSize).catch((error) => { | ||||||
| 			toast.error(error); | 			toast.error(error); | ||||||
| 			return null; | 			return null; | ||||||
|  | @ -119,6 +140,23 @@ | ||||||
| 		<div> | 		<div> | ||||||
| 			<div class=" mb-1 text-sm font-medium">{$i18n.t('Image Settings')}</div> | 			<div class=" mb-1 text-sm font-medium">{$i18n.t('Image Settings')}</div> | ||||||
| 
 | 
 | ||||||
|  | 			<div class=" py-0.5 flex w-full justify-between"> | ||||||
|  | 				<div class=" self-center text-xs font-medium">Image Generation Engine</div> | ||||||
|  | 				<div class="flex items-center relative"> | ||||||
|  | 					<select | ||||||
|  | 						class="w-fit pr-8 rounded px-2 p-1 text-xs bg-transparent outline-none text-right" | ||||||
|  | 						bind:value={imageGenerationEngine} | ||||||
|  | 						placeholder="Select a mode" | ||||||
|  | 						on:change={async () => { | ||||||
|  | 							await updateImageGeneration(); | ||||||
|  | 						}} | ||||||
|  | 					> | ||||||
|  | 						<option value="">Default (Automatic1111)</option> | ||||||
|  | 						<option value="openai">Open AI (Dall-E)</option> | ||||||
|  | 					</select> | ||||||
|  | 				</div> | ||||||
|  | 			</div> | ||||||
|  | 
 | ||||||
| 			<div> | 			<div> | ||||||
| 				<div class=" py-0.5 flex w-full justify-between"> | 				<div class=" py-0.5 flex w-full justify-between"> | ||||||
| 					<div class=" self-center text-xs font-medium"> | 					<div class=" self-center text-xs font-medium"> | ||||||
|  | @ -128,7 +166,17 @@ | ||||||
| 					<button | 					<button | ||||||
| 						class="p-1 px-3 text-xs flex rounded transition" | 						class="p-1 px-3 text-xs flex rounded transition" | ||||||
| 						on:click={() => { | 						on:click={() => { | ||||||
| 							toggleImageGeneration(); | 							if (imageGenerationEngine === '' && AUTOMATIC1111_BASE_URL === '') { | ||||||
|  | 								toast.error('AUTOMATIC1111 Base URL is required.'); | ||||||
|  | 								enableImageGeneration = false; | ||||||
|  | 							} else if (imageGenerationEngine === 'openai' && OPENAI_API_KEY === '') { | ||||||
|  | 								toast.error('OpenAI API Key is required.'); | ||||||
|  | 								enableImageGeneration = false; | ||||||
|  | 							} else { | ||||||
|  | 								enableImageGeneration = !enableImageGeneration; | ||||||
|  | 							} | ||||||
|  | 
 | ||||||
|  | 							updateImageGeneration(); | ||||||
| 						}} | 						}} | ||||||
| 						type="button" | 						type="button" | ||||||
| 					> | 					> | ||||||
|  | @ -143,50 +191,62 @@ | ||||||
| 		</div> | 		</div> | ||||||
| 		<hr class=" dark:border-gray-700" /> | 		<hr class=" dark:border-gray-700" /> | ||||||
| 
 | 
 | ||||||
| 		<div class=" mb-2.5 text-sm font-medium">{$i18n.t('AUTOMATIC1111 Base URL')}</div> | 		{#if imageGenerationEngine === ''} | ||||||
| 		<div class="flex w-full"> | 			<div class=" mb-2.5 text-sm font-medium">{$i18n.t('AUTOMATIC1111 Base URL')}</div> | ||||||
| 			<div class="flex-1 mr-2"> | 			<div class="flex w-full"> | ||||||
| 				<input | 				<div class="flex-1 mr-2"> | ||||||
| 					class="w-full rounded py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-800 outline-none" | 					<input | ||||||
| 					placeholder="Enter URL (e.g. http://127.0.0.1:7860/)" | 						class="w-full rounded-lg py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-850 outline-none" | ||||||
| 					bind:value={AUTOMATIC1111_BASE_URL} | 						placeholder="Enter URL (e.g. http://127.0.0.1:7860/)" | ||||||
| 				/> | 						bind:value={AUTOMATIC1111_BASE_URL} | ||||||
| 			</div> |  | ||||||
| 			<button |  | ||||||
| 				class="px-3 bg-gray-200 hover:bg-gray-300 dark:bg-gray-600 dark:hover:bg-gray-700 rounded transition" |  | ||||||
| 				type="button" |  | ||||||
| 				on:click={() => { |  | ||||||
| 					// updateOllamaAPIUrlHandler(); |  | ||||||
| 
 |  | ||||||
| 					updateAUTOMATIC1111UrlHandler(); |  | ||||||
| 				}} |  | ||||||
| 			> |  | ||||||
| 				<svg |  | ||||||
| 					xmlns="http://www.w3.org/2000/svg" |  | ||||||
| 					viewBox="0 0 20 20" |  | ||||||
| 					fill="currentColor" |  | ||||||
| 					class="w-4 h-4" |  | ||||||
| 				> |  | ||||||
| 					<path |  | ||||||
| 						fill-rule="evenodd" |  | ||||||
| 						d="M15.312 11.424a5.5 5.5 0 01-9.201 2.466l-.312-.311h2.433a.75.75 0 000-1.5H3.989a.75.75 0 00-.75.75v4.242a.75.75 0 001.5 0v-2.43l.31.31a7 7 0 0011.712-3.138.75.75 0 00-1.449-.39zm1.23-3.723a.75.75 0 00.219-.53V2.929a.75.75 0 00-1.5 0V5.36l-.31-.31A7 7 0 003.239 8.188a.75.75 0 101.448.389A5.5 5.5 0 0113.89 6.11l.311.31h-2.432a.75.75 0 000 1.5h4.243a.75.75 0 00.53-.219z" |  | ||||||
| 						clip-rule="evenodd" |  | ||||||
| 					/> | 					/> | ||||||
| 				</svg> | 				</div> | ||||||
| 			</button> | 				<button | ||||||
| 		</div> | 					class="px-3 bg-gray-200 hover:bg-gray-300 dark:bg-gray-600 dark:hover:bg-gray-700 rounded-lg transition" | ||||||
|  | 					type="button" | ||||||
|  | 					on:click={() => { | ||||||
|  | 						// updateOllamaAPIUrlHandler(); | ||||||
| 
 | 
 | ||||||
| 		<div class="mt-2 text-xs text-gray-400 dark:text-gray-500"> | 						updateAUTOMATIC1111UrlHandler(); | ||||||
| 			{$i18n.t('Include `--api` flag when running stable-diffusion-webui')} | 					}} | ||||||
| 			<a | 				> | ||||||
| 				class=" text-gray-300 font-medium underline" | 					<svg | ||||||
| 				href="https://github.com/AUTOMATIC1111/stable-diffusion-webui/discussions/3734" | 						xmlns="http://www.w3.org/2000/svg" | ||||||
| 				target="_blank" | 						viewBox="0 0 20 20" | ||||||
| 			> | 						fill="currentColor" | ||||||
| 				<br /> | 						class="w-4 h-4" | ||||||
| 				{$i18n.t('(e.g. `sh webui.sh --api`)')} | 					> | ||||||
| 			</a> | 						<path | ||||||
| 		</div> | 							fill-rule="evenodd" | ||||||
|  | 							d="M15.312 11.424a5.5 5.5 0 01-9.201 2.466l-.312-.311h2.433a.75.75 0 000-1.5H3.989a.75.75 0 00-.75.75v4.242a.75.75 0 001.5 0v-2.43l.31.31a7 7 0 0011.712-3.138.75.75 0 00-1.449-.39zm1.23-3.723a.75.75 0 00.219-.53V2.929a.75.75 0 00-1.5 0V5.36l-.31-.31A7 7 0 003.239 8.188a.75.75 0 101.448.389A5.5 5.5 0 0113.89 6.11l.311.31h-2.432a.75.75 0 000 1.5h4.243a.75.75 0 00.53-.219z" | ||||||
|  | 							clip-rule="evenodd" | ||||||
|  | 						/> | ||||||
|  | 					</svg> | ||||||
|  | 				</button> | ||||||
|  | 			</div> | ||||||
|  | 
 | ||||||
|  | 			<div class="mt-2 text-xs text-gray-400 dark:text-gray-500"> | ||||||
|  | 				Include `--api` flag when running stable-diffusion-webui | ||||||
|  | 				<a | ||||||
|  | 					class=" text-gray-300 font-medium" | ||||||
|  | 					href="https://github.com/AUTOMATIC1111/stable-diffusion-webui/discussions/3734" | ||||||
|  | 					target="_blank" | ||||||
|  | 				> | ||||||
|  | 					(e.g. `sh webui.sh --api`) | ||||||
|  | 				</a> | ||||||
|  | 			</div> | ||||||
|  | 		{:else if imageGenerationEngine === 'openai'} | ||||||
|  | 			<div class=" mb-2.5 text-sm font-medium">OpenAI API Key</div> | ||||||
|  | 			<div class="flex w-full"> | ||||||
|  | 				<div class="flex-1 mr-2"> | ||||||
|  | 					<input | ||||||
|  | 						class="w-full rounded-lg py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-850 outline-none" | ||||||
|  | 						placeholder="Enter API Key" | ||||||
|  | 						bind:value={OPENAI_API_KEY} | ||||||
|  | 					/> | ||||||
|  | 				</div> | ||||||
|  | 			</div> | ||||||
|  | 		{/if} | ||||||
| 
 | 
 | ||||||
| 		{#if enableImageGeneration} | 		{#if enableImageGeneration} | ||||||
| 			<hr class=" dark:border-gray-700" /> | 			<hr class=" dark:border-gray-700" /> | ||||||
|  | @ -196,7 +256,7 @@ | ||||||
| 				<div class="flex w-full"> | 				<div class="flex w-full"> | ||||||
| 					<div class="flex-1 mr-2"> | 					<div class="flex-1 mr-2"> | ||||||
| 						<select | 						<select | ||||||
| 							class="w-full rounded py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-800 outline-none" | 							class="w-full rounded-lg py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-850 outline-none" | ||||||
| 							bind:value={selectedModel} | 							bind:value={selectedModel} | ||||||
| 							placeholder={$i18n.t('Select a model')} | 							placeholder={$i18n.t('Select a model')} | ||||||
| 						> | 						> | ||||||
|  | @ -204,9 +264,7 @@ | ||||||
| 								<option value="" disabled selected>{$i18n.t('Select a model')}</option> | 								<option value="" disabled selected>{$i18n.t('Select a model')}</option> | ||||||
| 							{/if} | 							{/if} | ||||||
| 							{#each models ?? [] as model} | 							{#each models ?? [] as model} | ||||||
| 								<option value={model.title} class="bg-gray-100 dark:bg-gray-700" | 								<option value={model.id} class="bg-gray-100 dark:bg-gray-700">{model.name}</option> | ||||||
| 									>{model.model_name}</option |  | ||||||
| 								> |  | ||||||
| 							{/each} | 							{/each} | ||||||
| 						</select> | 						</select> | ||||||
| 					</div> | 					</div> | ||||||
|  | @ -218,7 +276,7 @@ | ||||||
| 				<div class="flex w-full"> | 				<div class="flex w-full"> | ||||||
| 					<div class="flex-1 mr-2"> | 					<div class="flex-1 mr-2"> | ||||||
| 						<input | 						<input | ||||||
| 							class="w-full rounded py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-800 outline-none" | 							class="w-full rounded-lg py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-850 outline-none" | ||||||
| 							placeholder="Enter Image Size (e.g. 512x512)" | 							placeholder="Enter Image Size (e.g. 512x512)" | ||||||
| 							bind:value={imageSize} | 							bind:value={imageSize} | ||||||
| 						/> | 						/> | ||||||
|  | @ -231,7 +289,7 @@ | ||||||
| 				<div class="flex w-full"> | 				<div class="flex w-full"> | ||||||
| 					<div class="flex-1 mr-2"> | 					<div class="flex-1 mr-2"> | ||||||
| 						<input | 						<input | ||||||
| 							class="w-full rounded py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-800 outline-none" | 							class="w-full rounded-lg py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-850 outline-none" | ||||||
| 							placeholder="Enter Number of Steps (e.g. 50)" | 							placeholder="Enter Number of Steps (e.g. 50)" | ||||||
| 							bind:value={steps} | 							bind:value={steps} | ||||||
| 						/> | 						/> | ||||||
|  |  | ||||||
|  | @ -29,6 +29,7 @@ | ||||||
| 	let liteLLMAPIBase = ''; | 	let liteLLMAPIBase = ''; | ||||||
| 	let liteLLMAPIKey = ''; | 	let liteLLMAPIKey = ''; | ||||||
| 	let liteLLMRPM = ''; | 	let liteLLMRPM = ''; | ||||||
|  | 	let liteLLMMaxTokens = ''; | ||||||
| 
 | 
 | ||||||
| 	let deleteLiteLLMModelId = ''; | 	let deleteLiteLLMModelId = ''; | ||||||
| 
 | 
 | ||||||
|  | @ -336,7 +337,8 @@ | ||||||
| 				model: liteLLMModel, | 				model: liteLLMModel, | ||||||
| 				api_base: liteLLMAPIBase, | 				api_base: liteLLMAPIBase, | ||||||
| 				api_key: liteLLMAPIKey, | 				api_key: liteLLMAPIKey, | ||||||
| 				rpm: liteLLMRPM | 				rpm: liteLLMRPM, | ||||||
|  | 				max_tokens: liteLLMMaxTokens | ||||||
| 			}).catch((error) => { | 			}).catch((error) => { | ||||||
| 				toast.error(error); | 				toast.error(error); | ||||||
| 				return null; | 				return null; | ||||||
|  | @ -356,6 +358,7 @@ | ||||||
| 		liteLLMAPIBase = ''; | 		liteLLMAPIBase = ''; | ||||||
| 		liteLLMAPIKey = ''; | 		liteLLMAPIKey = ''; | ||||||
| 		liteLLMRPM = ''; | 		liteLLMRPM = ''; | ||||||
|  | 		liteLLMMaxTokens = ''; | ||||||
| 
 | 
 | ||||||
| 		liteLLMModelInfo = await getLiteLLMModelInfo(localStorage.token); | 		liteLLMModelInfo = await getLiteLLMModelInfo(localStorage.token); | ||||||
| 		models.set(await getModels()); | 		models.set(await getModels()); | ||||||
|  | @ -838,6 +841,22 @@ | ||||||
| 									</div> | 									</div> | ||||||
| 								</div> | 								</div> | ||||||
| 							</div> | 							</div> | ||||||
|  | 
 | ||||||
|  | 							<div> | ||||||
|  | 								<div class="mb-1.5 text-sm font-medium">Max Tokens</div> | ||||||
|  | 								<div class="flex w-full"> | ||||||
|  | 									<div class="flex-1"> | ||||||
|  | 										<input | ||||||
|  | 											class="w-full rounded-lg py-2 px-4 text-sm dark:text-gray-300 dark:bg-gray-850 outline-none" | ||||||
|  | 											placeholder="Enter Max Tokens (litellm_params.max_tokens)" | ||||||
|  | 											bind:value={liteLLMMaxTokens} | ||||||
|  | 											type="number" | ||||||
|  | 											min="1" | ||||||
|  | 											autocomplete="off" | ||||||
|  | 										/> | ||||||
|  | 									</div> | ||||||
|  | 								</div> | ||||||
|  | 							</div> | ||||||
| 						{/if} | 						{/if} | ||||||
| 					</div> | 					</div> | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -1,18 +1,23 @@ | ||||||
| <script lang="ts"> | <script lang="ts"> | ||||||
|  | 	import { WEBUI_BASE_URL } from '$lib/constants'; | ||||||
| 	import ImagePreview from './ImagePreview.svelte'; | 	import ImagePreview from './ImagePreview.svelte'; | ||||||
| 
 | 
 | ||||||
| 	export let src = ''; | 	export let src = ''; | ||||||
| 	export let alt = ''; | 	export let alt = ''; | ||||||
| 
 | 
 | ||||||
|  | 	let _src = ''; | ||||||
|  | 
 | ||||||
|  | 	$: _src = src.startsWith('/') ? `${WEBUI_BASE_URL}${src}` : src; | ||||||
|  | 
 | ||||||
| 	let showImagePreview = false; | 	let showImagePreview = false; | ||||||
| </script> | </script> | ||||||
| 
 | 
 | ||||||
| <ImagePreview bind:show={showImagePreview} {src} {alt} /> | <ImagePreview bind:show={showImagePreview} src={_src} {alt} /> | ||||||
| <button | <button | ||||||
| 	on:click={() => { | 	on:click={() => { | ||||||
| 		console.log('image preview'); | 		console.log('image preview'); | ||||||
| 		showImagePreview = true; | 		showImagePreview = true; | ||||||
| 	}} | 	}} | ||||||
| > | > | ||||||
| 	<img {src} {alt} class=" max-h-96 rounded-lg" draggable="false" /> | 	<img src={_src} {alt} class=" max-h-96 rounded-lg" draggable="false" /> | ||||||
| </button> | </button> | ||||||
|  |  | ||||||
|  | @ -64,12 +64,16 @@ | ||||||
| 	}; | 	}; | ||||||
| 
 | 
 | ||||||
| 	const editChatTitle = async (id, _title) => { | 	const editChatTitle = async (id, _title) => { | ||||||
| 		title = _title; | 		if (_title === '') { | ||||||
|  | 			toast.error('Title cannot be an empty string.'); | ||||||
|  | 		} else { | ||||||
|  | 			title = _title; | ||||||
| 
 | 
 | ||||||
| 		await updateChatById(localStorage.token, id, { | 			await updateChatById(localStorage.token, id, { | ||||||
| 			title: _title | 				title: _title | ||||||
| 		}); | 			}); | ||||||
| 		await chats.set(await getChatList(localStorage.token)); | 			await chats.set(await getChatList(localStorage.token)); | ||||||
|  | 		} | ||||||
| 	}; | 	}; | ||||||
| 
 | 
 | ||||||
| 	const deleteChat = async (id) => { | 	const deleteChat = async (id) => { | ||||||
|  | @ -393,12 +397,13 @@ | ||||||
| 										show = false; | 										show = false; | ||||||
| 									} | 									} | ||||||
| 								}} | 								}} | ||||||
|  | 								draggable="false" | ||||||
| 							> | 							> | ||||||
| 								<div class=" flex self-center flex-1 w-full"> | 								<div class=" flex self-center flex-1 w-full"> | ||||||
| 									<div | 									<div | ||||||
| 										class=" text-left self-center overflow-hidden {chat.id === $chatId | 										class=" text-left self-center overflow-hidden {chat.id === $chatId | ||||||
| 											? 'w-[160px]' | 											? 'w-[160px]' | ||||||
| 											: 'w-full'} " | 											: 'w-full'}  h-[20px]" | ||||||
| 									> | 									> | ||||||
| 										{chat.title} | 										{chat.title} | ||||||
| 									</div> | 									</div> | ||||||
|  |  | ||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue