Compare commits

..

328 Commits

Author SHA1 Message Date
vegu-ai-tools
65a8cbf853 linting 2025-10-24 18:02:48 +03:00
vegu-ai-tools
7dbff08056 restore input focus after autocomplete 2025-10-24 18:02:20 +03:00
vegu-ai-tools
159efc500d improve configuration issue alert visibility 2025-10-24 17:56:38 +03:00
vegu-ai-tools
c0a11b8546 fix some errors during kcpp client deletion 2025-10-24 17:46:42 +03:00
vegu-ai-tools
f90099999c nothing to detemrine of no model is sent 2025-10-24 17:46:28 +03:00
vegu-ai-tools
3ca4299214 linting 2025-10-23 19:02:49 +03:00
vegu-ai-tools
260f9fa374 enhance error logging in background processing to include traceback information 2025-10-23 19:02:08 +03:00
vegu-ai-tools
bed6158003 remove debug cruft 2025-10-23 01:47:53 +03:00
vegu-ai-tools
bc5234a3a4 linting 2025-10-21 23:55:14 +03:00
vegu-ai-tools
ca1f761ad7 separate message processing from main loop 2025-10-21 23:53:05 +03:00
vegu-ai-tools
1422fc4541 unhandled errors at the loop level should not crash the entire scene 2025-10-21 23:46:01 +03:00
vegu-ai-tools
b71d9a580f formatting fixes 2025-10-21 02:10:18 +03:00
vegu-ai-tools
248333139a increase font size 2025-10-21 02:01:58 +03:00
vegu-ai-tools
fb4e73c6e8 fix issue where cancelling some generations would cause errors 2025-10-21 01:54:57 +03:00
vegu-ai-tools
ae36f8491c removing base attrib ute or detail also clears it from shared list 2025-10-21 01:48:08 +03:00
vegu-ai-tools
501668b2fe docs 2025-10-19 23:04:01 +03:00
vegu-ai-tools
cbcef92ed7 linting 2025-10-19 22:49:15 +03:00
vegu-ai-tools
8a95b02099 : in world entry titles will now load correctly 2025-10-19 22:49:01 +03:00
vegu-ai-tools
5c33723b7b docs 2025-10-19 22:04:07 +03:00
vegu-ai-tools
1ab24396f1 summarizer fire off of push_history.after 2025-10-19 18:49:56 +03:00
vegu-ai-tools
b6729f3290 tweak defaults 2025-10-19 18:31:26 +03:00
vegu-ai-tools
913db13590 linting 2025-10-19 17:32:30 +03:00
vegu-ai-tools
ba1e64d359 only allow forking on saved messages 2025-10-19 17:32:14 +03:00
vegu-ai-tools
519b600bc9 Update RequestInput.vue to handle extra_params more robustly, ensuring defaults are set correctly for input. 2025-10-19 17:31:44 +03:00
vegu-ai-tools
9901c36af6 emit_status export rev 2025-10-19 17:29:27 +03:00
vegu-ai-tools
6d4bfd59ac forked scenes reset memory id and are not immutable 2025-10-19 17:17:37 +03:00
vegu-ai-tools
466bac8061 Refactor scene reference handling in delete_changelog_files to prevent incorrect deletions. Added a test to verify proper scene reference construction and ensure changelog files are deleted correctly. 2025-10-19 17:17:20 +03:00
vegu-ai-tools
6328062c3d gracefully handle removed attributes 2025-10-19 17:10:37 +03:00
vegu-ai-tools
f9c1228b3e linting 2025-10-19 15:26:50 +03:00
vegu-ai-tools
5d40b650dc prompt tweaks 2025-10-19 15:26:39 +03:00
vegu-ai-tools
aeef4c266f improve autocomplete handling when prefill isn't available 2025-10-19 15:20:34 +03:00
vegu-ai-tools
e7180b0dd5 fix issue where fork / restore would restore duplicate messages 2025-10-19 14:45:01 +03:00
vegu-ai-tools
154a02adf0 opse 4.5 isnt a thing 2025-10-17 19:30:34 +03:00
vegu-ai-tools
fe73970f67 add heiku 4.5 model and make default 2025-10-17 19:26:42 +03:00
vegu-ai-tools
007b944c4a linting 2025-10-17 18:32:21 +03:00
vegu-ai-tools
8a79edc693 Update default_player_character assignment in ConfigPlugin to use GamePlayerCharacter schema for improved data validation 2025-10-17 18:32:05 +03:00
vegu-ai-tools
1e29c7eab4 fix issue where valid data processed in extract_data_with_ai_fallback was not returned 2025-10-15 02:20:10 +03:00
vegu-ai-tools
07f1a72618 Add field validator for lock_template in Client model to ensure boolean value is returned 2025-10-14 01:37:22 +03:00
vegu-ai-tools
d51c7a2700 Refactor lock_template field in Client model and ClientModal component to ensure consistent boolean handling 2025-10-14 01:34:11 +03:00
vegu-ai-tools
bb14f90669 Remove unused template_file field from Defaults model in Client configuration 2025-10-14 01:29:50 +03:00
vegu-ai-tools
4791918e34 Update lock_template field in Client model to allow None type in addition to bool 2025-10-14 01:26:01 +03:00
vegu-ai-tools
5d6a4eef63 docs 2025-10-14 00:55:58 +03:00
vegu-ai-tools
875deb2682 Update CharacterContextItem to allow 'value' to accept dict type in addition to existing types 2025-10-14 00:10:18 +03:00
vegu-ai-tools
cc60ee1beb fix direct_narrator character argument 2025-10-13 23:48:22 +03:00
vegu-ai-tools
db8c021b68 There is no longer a point to enforcing creative mode when there are no characters 2025-10-13 23:26:51 +03:00
vegu-ai-tools
7d7f210a2f persist client template lock through model changes 2025-10-13 12:01:33 +03:00
vegu-ai-tools
27bc0a5b2f linting 2025-10-13 00:41:13 +03:00
vegu-ai-tools
d25fdd9422 Add lock_template feature to Client configuration and update related components for template management 2025-10-13 00:40:58 +03:00
vegu-ai-tools
f5a2b9a67b Add TODO comments in finalize_llama3 and finalize_YI methods to indicate removable cruft 2025-10-12 23:26:01 +03:00
vegu-ai-tools
8ebcd4ba5d fix crash when attempting to delete some clients 2025-10-12 22:51:29 +03:00
vegu-ai-tools
ffc6d75f46 Refine agent persona description in WorldStateManagerTemplates to clarify assignment per agent in Scene Settings, maintaining focus on current director-only support. 2025-10-12 18:11:06 +03:00
vegu-ai-tools
160b9c4e69 Update agent persona description in WorldStateManagerTemplates to specify current support for director only, enhancing clarity for users. 2025-10-12 18:08:47 +03:00
vegu-ai-tools
2e925c95b5 Enhance NewSceneSetupModal to include subtitles for writing styles and director personas, improving context and usability. 2025-10-12 18:08:40 +03:00
vegu-ai-tools
7edb830246 docs 2025-10-12 18:04:07 +03:00
vegu-ai-tools
61bde467cd Enhance NodeEditorLibrary by adding primary color to tree component for improved visibility and user experience. 2025-10-12 18:03:53 +03:00
vegu-ai-tools
1274155e84 docs 2025-10-12 17:06:02 +03:00
vegu-ai-tools
7569998cdb Add support for project-specific grouping in NodeEditorLibrary for templates/modules, enhancing organization of node groups. 2025-10-12 17:05:11 +03:00
vegu-ai-tools
a9fed9a4dd Add Nexus agent persona to talemate template and initialize phrases array 2025-10-12 14:51:41 +03:00
vegu-ai-tools
7f0c8a339e 0.33 added 2025-10-12 01:14:00 +03:00
vegu-ai-tools
2cc1c8c6ed director action module updates 2025-10-11 19:31:40 +03:00
vegu-ai-tools
ce03243ecf linting 2025-10-11 19:31:08 +03:00
vegu-ai-tools
1923282727 Increase maximum changelog file size limit from 500KB to 1MB to accommodate larger change logs. 2025-10-11 19:28:47 +03:00
vegu-ai-tools
0b41f76b07 Update EmitWorldEditorSync node to include websocket passthrough in sync action for improved event handling. 2025-10-11 19:18:13 +03:00
vegu-ai-tools
5d9fdca5d8 Update card styles in IntroRecentScenes.vue for improved visual consistency; change card color to grey-darken-3 and adjust text classes for titles and subtitles. 2025-10-11 19:17:47 +03:00
vegu-ai-tools
0b1a2c5159 Remove debug diagnostics from DirectorConsoleChats.vue to clean up console output. 2025-10-11 19:17:42 +03:00
vegu-ai-tools
b9dbe9179c Update usageCheatSheet in DirectorConsoleChatsToolbar.vue to include recommendation for 100B+ models. 2025-10-11 19:17:36 +03:00
vegu-ai-tools
9ec55a0004 director action updates 2025-10-11 14:08:44 +03:00
vegu-ai-tools
7529726df9 direct context update fn 2025-10-11 13:52:59 +03:00
vegu-ai-tools
ebc2ae804b director action module updates 2025-10-11 13:51:47 +03:00
vegu-ai-tools
2b3cb8d101 Update Anthropic client with new models and adjust default settings; introduce limited parameter models for specific configurations. 2025-10-11 13:51:22 +03:00
vegu-ai-tools
6c3b53e37c Add EmitWorldEditorSync node to handle world editor synchronization; update WorldStateManager to refresh active tab on sync action. 2025-10-11 13:51:14 +03:00
vegu-ai-tools
3dee0ec0e9 relock 2025-10-11 12:39:11 +03:00
vegu-ai-tools
504707796c linting 2025-10-10 16:28:44 +03:00
vegu-ai-tools
0f59e1cd21 Add cover image and writing style sections to story and character templates; update chat common tasks with new scene restrictions and user guide reference. 2025-10-10 16:28:34 +03:00
vegu-ai-tools
2f742b95d6 Update usageCheatSheet text in DirectorConsoleChatsToolbar.vue for clarity and add pre-wrap styling to tooltip 2025-10-10 16:18:00 +03:00
vegu-ai-tools
2627972d8b Refactor toggleNavigation method to accept an 'open' parameter for direct control over drawer visibility in TalemateApp.vue 2025-10-10 16:03:33 +03:00
vegu-ai-tools
934e62dded Add building blocks template for story configuration and scene management 2025-10-10 16:03:24 +03:00
vegu-ai-tools
238ff1dfe5 fix update_introduction 2025-10-10 15:46:44 +03:00
vegu-ai-tools
8b4e1962c4 nodes 2025-10-10 15:43:18 +03:00
vegu-ai-tools
d3a9c7f2c1 fix tests 2025-10-09 04:33:07 +03:00
vegu-ai-tools
b8df7cfed8 linting 2025-10-09 04:17:07 +03:00
vegu-ai-tools
4b27917173 fix issue with data structure parsing 2025-10-09 04:16:17 +03:00
vegu-ai-tools
8b94312f1f Add MAX_CONTENT_WIDTH constant and update components to use it for consistent max width styling 2025-10-09 03:13:52 +03:00
vegu-ai-tools
15aea906c5 Update SharedContext to use await for set_shared method, ensuring proper asynchronous handling when modifying character sharing status. 2025-10-09 02:51:32 +03:00
vegu-ai-tools
fe36431a27 Refactor WorldStateManagerSceneSharedContext.vue to improve cancel functionality by introducing a dedicated cancelCreate method and removing the direct dialog toggle from the Cancel button. This enhances code clarity and maintainability. 2025-10-09 02:51:28 +03:00
vegu-ai-tools
f6dabc18eb Add intent_state to SceneInitialization model and update load_scene_from_data function to handle intent state. Introduce story_intent property in Scene class and reset method in SceneIntent class. Update WorldStateManagerSceneSharedContext.vue to include intent state in scene initialization parameters. 2025-10-09 02:41:01 +03:00
vegu-ai-tools
b5e30600fa Refactor CoverImage component to enhance drag-and-drop functionality and improve styling for empty portrait state. 2025-10-09 01:56:42 +03:00
vegu-ai-tools
765cd5799c Add assets field to SceneInitialization model and update load_scene_from_data function to handle scene assets. Update WorldStateManagerSceneSharedContext.vue to include assets in scene initialization parameters. 2025-10-09 01:45:13 +03:00
vegu-ai-tools
9457620767 linting 2025-10-08 00:37:30 +03:00
vegu-ai-tools
8b07422939 nodes updated 2025-10-07 03:33:57 +03:00
vegu-ai-tools
4eb185e895 Add CreateStaticArchiveEntry and RemoveStaticArchiveEntry nodes for managing static history entries. Implement input/output properties and error handling for entry creation and deletion. 2025-10-07 03:33:16 +03:00
vegu-ai-tools
78d434afd3 Add "static history" option to ContextualGenerate node for enhanced contextual generation capabilities. 2025-10-07 03:33:10 +03:00
vegu-ai-tools
dc2b1c9149 Add is_static property to HistoryEntry for static history entry identification 2025-10-07 03:33:01 +03:00
vegu-ai-tools
511a33f69f allow contextual generation of static history entries 2025-10-07 03:32:52 +03:00
vegu-ai-tools
0e3eb15fce linting 2025-10-05 03:29:33 +03:00
vegu-ai-tools
f925758319 Refactor response identifier in RevisionMixin to dynamically use calculated response length for improved prompt handling. 2025-10-05 03:27:27 +03:00
vegu-ai-tools
81e1da3c21 Update response length calculation in RevisionMixin to include token count for improved text processing. 2025-10-05 03:25:02 +03:00
vegu-ai-tools
9dea2daef5 Enhance NarratorAgent to support dynamic response length configuration. Updated max generation length from 192 to 256 tokens and introduced a new method to calculate response length. Modified narration methods to accept and utilize response length parameter. Added response length property in GenerateNarrationBase class and updated templates to include response length handling. 2025-10-05 03:24:39 +03:00
vegu-ai-tools
1c592a438f Add Seed.jinja2 template for LLM prompts with reasoning patterns and user interaction handling 2025-10-05 03:10:44 +03:00
vegu-ai-tools
fadf4b8f2d allow prompt templates to specify reasoning pattern 2025-10-05 03:10:31 +03:00
vegu-ai-tools
22f97f60ea Update GLM-no-reasoning template to include <think></think> tag before coercion message for improved prompt structure. 2025-10-05 01:52:32 +03:00
vegu-ai-tools
6782bfe93f Change log level from warning to debug for migrate_narrator_source_to_meta error handling in NarratorMessage class. 2025-10-04 23:18:07 +03:00
vegu-ai-tools
2c2d2f160c linting 2025-10-04 22:35:06 +03:00
vegu-ai-tools
3b85d007e2 Refactor WorldStateManager components to enhance history management and sharing capabilities. Added summarized history titles, improved UI for sharing static history, and integrated scene summarization functionality. Removed deprecated methods related to shared context settings. 2025-10-04 22:29:43 +03:00
vegu-ai-tools
ca1a1872ec Update icon for AgentWebsocketHandler in NodeEditorLibrary component to mdi-web-box 2025-10-04 22:29:32 +03:00
vegu-ai-tools
1f5baf9958 Update scene loading to allow setting scene ID from data and include ID in scene serialization 2025-10-04 22:29:22 +03:00
vegu-ai-tools
b9dcfd54a5 Enhance GetWorldEntry node to include 'shared' property in output values from world entry context 2025-10-04 22:29:08 +03:00
vegu-ai-tools
ceb998088f Update manual context handling in WorldStateManager to include shared property from existing context 2025-10-04 22:28:55 +03:00
vegu-ai-tools
a21117674b Add data property to QueueResponse class for websocket communication and update run method to include action and data in output values. 2025-10-04 22:22:39 +03:00
vegu-ai-tools
40bdf7b361 nodes 2025-10-04 22:22:26 +03:00
vegu-ai-tools
7161786b50 Add SummarizeWebsocketHandler to handle summarize actions and integrate it into SummarizeAgent 2025-10-04 20:01:18 +03:00
vegu-ai-tools
5c960784d9 Add check for node selectability in NodeEditorNodeSearch component to filter search results accordingly. 2025-10-04 20:01:06 +03:00
vegu-ai-tools
0fd6d01184 Add Agent Websocket Handler option to Node Editor Library with corresponding icons and labels 2025-10-04 20:01:01 +03:00
vegu-ai-tools
4a9522d030 Add characters output to ContextHistory node to track active participants in the scene 2025-10-04 20:00:52 +03:00
vegu-ai-tools
3a00e33dc1 Refactor init_nodes method in DirectorAgent to call superclass method and rename chat initialization method in DirectorChatMixin for clarity. 2025-10-04 20:00:43 +03:00
vegu-ai-tools
751a2acfcb agent websocket handler node support 2025-10-04 20:00:32 +03:00
vegu-ai-tools
acb3b66328 Add active frontend websocket handler management in websocket_endpoint 2025-10-04 19:59:07 +03:00
vegu-ai-tools
8d56eb1ff8 Update TalemateApp.vue to set the active tab to 'main' when switching to the node editor, improving navigation consistency. 2025-10-04 19:58:39 +03:00
vegu-ai-tools
435082935e prompt tweaks 2025-10-04 16:39:24 +03:00
vegu-ai-tools
d7fe0e36d9 prompt tweaks 2025-10-04 16:07:38 +03:00
vegu-ai-tools
0ae92c39de Implement logic to always show scene view in scene mode within TalemateApp.vue, enhancing user experience during scene interactions. 2025-10-04 15:41:19 +03:00
vegu-ai-tools
cfc84eb357 Refactor source entry attribute access in collect_source_entries function to use getattr for optional attributes, improving robustness. 2025-10-04 15:20:23 +03:00
vegu-ai-tools
0336f19066 linting 2025-10-04 14:42:25 +03:00
vegu-ai-tools
e2cca100f5 Update IntroRecentScenes.vue to use optional chaining for selectedScene properties and enhance backup timestamp display with revision info. 2025-10-04 14:41:50 +03:00
vegu-ai-tools
51d4ae57e9 Skip processing of changelog files in _list_files_and_directories function to prevent unnecessary inclusion in file listings. 2025-10-04 14:41:34 +03:00
vegu-ai-tools
8714fd1726 Update _apply_delta function to enhance delta application handling by adding parameters for error logging and force application of changes on non-existent paths. 2025-10-04 14:41:25 +03:00
vegu-ai-tools
7e4d3b7268 Enhance backup restore functionality by adding base and latest snapshot options; improve UI with clearer labels and alerts for restore actions. 2025-10-04 14:09:59 +03:00
vegu-ai-tools
117ea78f06 Add ensure_changelogs_for_all_scenes function to manage changelog files for all scenes; integrate it into the server run process. 2025-10-04 13:42:08 +03:00
vegu-ai-tools
5f396b22d2 Add update_from_scene method calls in SharedContextMixin for scene synchronization 2025-10-04 00:47:16 +03:00
vegu-ai-tools
430c69a83d Refactor character removal logic in shared context to prevent deletion; characters are now only marked as non-shared. 2025-10-04 00:43:23 +03:00
vegu-ai-tools
9d38432a8b avoid changed size error 2025-10-04 00:21:11 +03:00
vegu-ai-tools
bd378f4f44 missing arg 2025-10-04 00:17:35 +03:00
vegu-ai-tools
e800179a0c activate needs to happen explicitly now and deactivated is the default 2025-10-04 00:15:08 +03:00
vegu-ai-tools
a21c3d2ccf properly activate characters 2025-10-04 00:10:48 +03:00
vegu-ai-tools
d41953a70d linting 2025-10-03 23:46:10 +03:00
vegu-ai-tools
99efbade54 prompt tweaks 2025-10-03 23:45:20 +03:00
vegu-ai-tools
6db3cb72ff ensure character gets added to character_data 2025-10-03 23:45:15 +03:00
vegu-ai-tools
68ed364270 relock 2025-10-03 22:31:03 +03:00
vegu-ai-tools
665cc6f4b1 Refactor base_attributes type in Character model to a more generic dict type for improved flexibility 2025-10-03 22:29:59 +03:00
vegu-ai-tools
7ef5b70ff2 typo 2025-10-03 22:29:40 +03:00
vegu-ai-tools
a6d1065dcb Improve error handling in export_node_definitions by adding a try-except block for module path resolution. Log a warning if the relative path conversion fails. 2025-10-03 15:54:24 +03:00
vegu-ai-tools
33d093e5bb show icons 2025-10-03 15:52:19 +03:00
vegu-ai-tools
b2035ddebe Add relative_to_root function for path resolution and update node export logic
- Introduced a new function `relative_to_root` in path.py to resolve paths relative to the TALEMATE_ROOT.
- Updated the `export_node_definitions` function in registry.py to use `relative_to_root` for module path resolution.
- Added a check to skip non-selectable node definitions in litegraphUtils.js during registration.
2025-10-03 15:51:21 +03:00
vegu-ai-tools
3e06f8a64d node fixes 2025-10-03 15:07:39 +03:00
vegu-ai-tools
152fb9b474 linting 2025-10-03 14:48:26 +03:00
vegu-ai-tools
b7fdc3dc38 Enhance data extraction in Focal class by adding a fallback mechanism. Implemented additional error handling to attempt data extraction from a fenced block if the initial extraction fails, improving robustness in handling responses. 2025-10-03 14:47:49 +03:00
vegu-ai-tools
3cf5557780 prompt tweaks 2025-10-03 14:47:36 +03:00
vegu-ai-tools
e29496f650 recet scene message visibility on scene load 2025-10-03 14:47:30 +03:00
vegu-ai-tools
2bb1e45eb5 Enhance error handling in DynamicInstruction class by enforcing header requirement and ensuring content defaults to an empty string if not provided. 2025-10-03 14:47:07 +03:00
vegu-ai-tools
4dc580f630 remove debug msg 2025-10-03 14:46:58 +03:00
vegu-ai-tools
6a27c49594 node fixes 2025-10-03 14:45:57 +03:00
vegu-ai-tools
131479e29b Add chat template identifier support and error handling in ModelPrompt class
- Implemented logic to check for 'chat_template.jinja2' in Hugging Face repository.
- Added new template identifiers: GraniteIdentifier and GLMIdentifier.
- Enhanced error handling to avoid logging 404 errors for missing templates.
- Introduced Granite.jinja2 template file for prompt structure.
2025-10-03 14:45:46 +03:00
vegu-ai-tools
e6d528323b Remove plan.md 2025-10-03 12:39:18 +03:00
vegu-ai-tools
ccf6284442 linting 2025-10-02 15:41:07 +03:00
vegu-ai-tools
99a9488564 Update clear chat button logic to consider appBusy state in DirectorConsoleChatsToolbar component, enhancing user experience during busy operations. 2025-10-02 15:40:43 +03:00
vegu-ai-tools
42d08e5ac9 Refactor DirectorChatMixin to utilize standalone utility functions for parsing response sections and extracting action blocks. This improves code clarity and maintainability. Added tests for new utility functions in test_utils_prompt.py to ensure correct functionality. 2025-10-02 15:40:31 +03:00
vegu-ai-tools
b137546697 Add appBusy prop to DirectorConsoleChats and DirectorConsoleChatsToolbar components to manage button states during busy operations. 2025-10-02 15:37:37 +03:00
vegu-ai-tools
f13d306470 node updates 2025-10-02 14:48:30 +03:00
vegu-ai-tools
7aa274a0e0 linting 2025-10-02 14:44:38 +03:00
vegu-ai-tools
160818a26c Refactor ConfirmActionInline component to improve button rendering logic. Introduced 'size' prop for button customization and added 'comfortable' density option. Simplified icon handling with computed property for better clarity. 2025-10-02 14:44:26 +03:00
vegu-ai-tools
6953ccec69 director chat support remove message and regenerate message 2025-10-02 14:44:21 +03:00
vegu-ai-tools
b746b8773b Update EXTERNAL_DESCRIPTION in TabbyAPI client to include notes on EXL3 model sensitivity to inference parameters. Adjust handling of 'repetition_penalty_range' in parameter list for clarity. 2025-10-02 11:13:35 +03:00
vegu-ai-tools
8652e88ea8 Remove redundant question handling logic in DirectorChatMixin to streamline action selection process. 2025-10-01 22:45:56 +03:00
vegu-ai-tools
ad31e54e3a linting 2025-10-01 14:01:47 +03:00
vegu-ai-tools
748a2cfccf allow individual sharing of attributes and details 2025-10-01 13:54:08 +03:00
vegu-ai-tools
20db574155 Remove status emission for gameplay switch in CmdSetEnvironmentToScene class. 2025-10-01 10:32:37 +03:00
vegu-ai-tools
65e17f234f Add lastLoadedJSON property to GameState component for change detection. Update validation logic to prevent unnecessary updates when game state has not changed. 2025-10-01 10:29:41 +03:00
vegu-ai-tools
850679a0e8 Refactor GameState component to integrate Codemirror for JSON editing, replacing the previous treeview structure. Implement validation for JSON input and enhance error handling. Remove unused methods and streamline state management. 2025-10-01 10:27:13 +03:00
vegu-ai-tools
ae0749d173 linting 2025-10-01 02:41:35 +03:00
vegu-ai-tools
b13eb5be69 Add scene title generation to load process and update contextual generation template. Introduced a new method in AssistantMixin for generating scene titles, ensuring titles are concise and free of special characters. Updated load_scene_from_data to assign generated titles to scenes. 2025-10-01 02:41:26 +03:00
vegu-ai-tools
a4985c1888 Refine messages for shared context checkboxes in WorldStateManagerCharacter and WorldStateManagerWorldEntries components for clarity. 2025-10-01 02:41:14 +03:00
vegu-ai-tools
67571ec9be Update WorldStateManagerSceneSharedContext.vue to conditionally display alert based on scene saving status and new scene creation state. 2025-10-01 02:41:07 +03:00
vegu-ai-tools
29a5a5ac52 linting 2025-10-01 01:59:46 +03:00
vegu-ai-tools
60f5f20715 rename inheritance to scene initialization 2025-10-01 01:59:37 +03:00
vegu-ai-tools
00e3aa4a19 Add active_characters and intro_instructions to Inheritance model; implement intro generation in load_scene_from_data. Update WorldStateManagerSceneSharedContext.vue to enhance new scene creation dialog with character selection and premise instructions. 2025-10-01 01:44:26 +03:00
vegu-ai-tools
b59c3ab273 linting 2025-10-01 00:35:09 +03:00
vegu-ai-tools
4dbc824c07 Comment out 'repetition_penalty_range' in TabbyAPIClient to prevent unexpected "<unk><unk> .." responses. Further investigation needed. 2025-10-01 00:34:37 +03:00
vegu-ai-tools
0463fd37e5 Enhance chat modes by adding 'nospoilers' option to DirectorChat and related payloads. Update chat instructions to reflect new mode behavior and improve UI to support mode-specific icons and colors in the DirectorConsoleChatsToolbar. 2025-09-30 19:56:35 +03:00
vegu-ai-tools
a18a43cbe6 linting 2025-09-30 19:18:35 +03:00
vegu-ai-tools
b1cbaf650f Update WorldStateManagerSceneSharedContext.vue to clarify sharing of character, world entries, and history across connected scenes. 2025-09-30 19:18:22 +03:00
vegu-ai-tools
73c544211a shared context static history support
fix context memory db imports to always import
2025-09-30 19:15:29 +03:00
vegu-ai-tools
c65a7889d3 Enhance SharedContext.update_to_scene method to properly add or update character data in the scene based on existence checks. This improves the synchronization of character states between shared context and scene. 2025-09-30 14:08:34 +03:00
vegu-ai-tools
dc9e297587 Character.update deserialize voice value correctly 2025-09-30 14:08:19 +03:00
vegu-ai-tools
5d361331d5 comment 2025-09-30 14:07:48 +03:00
vegu-ai-tools
0b8810073f Refactor NodeEditor and TalemateApp components to enhance UI interactions. Removed the exit creative mode button from NodeEditor and updated tooltips for clarity. Adjusted app bar navigation icons for better accessibility and added functionality to switch between node editor and creative mode. 2025-09-30 13:44:30 +03:00
vegu-ai-tools
6ba65ff75e Refactor NodeEditorLibrary to improve search functionality and debounce input handling. Updated v-text-field model and added a watcher for search input to enhance performance. 2025-09-30 13:22:18 +03:00
veguAI
291921a9f2 Shared context 2 (#19)
Shared context
2025-09-30 03:26:48 +03:00
vegu-ai-tools
1b8ba12e61 fix world editor auto sync 2025-09-28 23:17:51 +03:00
vegu-ai-tools
9adbb2c518 fix button 2025-09-28 23:17:42 +03:00
vegu-ai-tools
883dffdd73 store character data at unified point 2025-09-28 22:41:17 +03:00
vegu-ai-tools
0c5fd2e48d Enhance DirectorConsoleChatsToolbar by adding a usage cheat sheet tooltip for user guidance and refining the Clear Chat button's UI for better accessibility. 2025-09-28 15:56:26 +03:00
vegu-ai-tools
7a6ae0f135 Update chat instructions to clarify user intent considerations and enhance decisiveness in responses. Added guidance on distinguishing between scene progression and background changes, and refined analysis requirements for user interactions. 2025-09-28 15:56:19 +03:00
vegu-ai-tools
6da7b29b94 Refactor push_history method to be asynchronous across multiple agents and scenes, ensuring consistent handling of message history updates. 2025-09-28 15:13:45 +03:00
vegu-ai-tools
b423bc3a18 Add scene progression guidance to chat-common-tasks template 2025-09-28 15:13:32 +03:00
vegu-ai-tools
bf8a580c33 relock 2025-09-28 14:41:55 +03:00
vegu-ai-tools
734b2bab19 linting 2025-09-28 14:39:27 +03:00
vegu-ai-tools
46afdeeb0b responsive layout fixes in template editors 2025-09-28 14:39:00 +03:00
vegu-ai-tools
674dfc5978 anchor clear chat confirm to top 2025-09-28 14:34:09 +03:00
vegu-ai-tools
71595a1fff Enhance ConfirmActionPrompt component by adding anchorTop prop for dynamic alignment and adjusting icon size and color for improved UI consistency. 2025-09-28 14:33:52 +03:00
vegu-ai-tools
219b5e2786 Enhance action handling in DirectorChatMixin by skipping actions when a question is present in the parsed response, ensuring better response accuracy. 2025-09-28 14:26:00 +03:00
vegu-ai-tools
1243d03718 director summary return appropriately on no action taken 2025-09-28 13:40:27 +03:00
vegu-ai-tools
ac9c66915b prompt tweaks 2025-09-28 05:11:25 +03:00
vegu-ai-tools
00b3c05f3d node updates 2025-09-28 05:03:24 +03:00
vegu-ai-tools
308363c93c node updates 2025-09-28 04:55:28 +03:00
vegu-ai-tools
78cc9334d3 node updates 2025-09-28 04:24:39 +03:00
vegu-ai-tools
e74e9c679a Add data_expected attribute to Focal and Prompt classes for enhanced response handling 2025-09-28 04:24:22 +03:00
vegu-ai-tools
6549d65ee8 linting 2025-09-28 03:51:29 +03:00
vegu-ai-tools
babe77929c node adjustments 2025-09-28 03:51:15 +03:00
vegu-ai-tools
531c0b4e87 prompt tweaks 2025-09-28 03:51:10 +03:00
vegu-ai-tools
3dc2269678 Add additional outputs for context validation in ValidateContextIDItem node, including context type, context value, and name. 2025-09-28 03:51:02 +03:00
vegu-ai-tools
26d7886c31 Add string replacement functionality and Jinja2 formatting support in nodes. Introduced 'old' and 'new' properties for substring replacement in the Replace node, and added a new Jinja2Format node for template rendering using jinja2. 2025-09-28 03:50:52 +03:00
vegu-ai-tools
3dadf49a69 Add context type output and filtering for creative context ID meta entries in PathToContextID and ContextIDMetaEntries nodes 2025-09-28 03:50:39 +03:00
vegu-ai-tools
4cb612bc23 prompt tweaks 2025-09-28 01:08:30 +03:00
vegu-ai-tools
b85b983522 node updates 2025-09-28 01:08:12 +03:00
vegu-ai-tools
922f520ec3 linting 2025-09-27 16:07:20 +03:00
vegu-ai-tools
555d90e53a immutable scenes should reset context db on load 2025-09-27 16:07:07 +03:00
vegu-ai-tools
e84c36a31b Enhance scene view toggle functionality to support shift-click behavior for closing all drawers when hiding the scene view. 2025-09-27 16:00:00 +03:00
vegu-ai-tools
072cd7fd12 linting 2025-09-27 15:53:37 +03:00
vegu-ai-tools
b9f5423f92 gamestate nodes 2025-09-27 15:52:17 +03:00
vegu-ai-tools
e9f0e4124a Add UnpackGameState node to retrieve and unpack game state variables 2025-09-27 15:52:07 +03:00
vegu-ai-tools
c361f4723b Add DictUpdate node 2025-09-27 15:51:25 +03:00
vegu-ai-tools
e0c92be628 Add 'data_multiple' property to GenerateResponse class to allow multiple data structures in responses. Update output socket type for 'data_obj' to support both dict and list formats. 2025-09-27 14:38:47 +03:00
vegu-ai-tools
52f07b26fa Refactor Prompt class by removing LoopedPrompt and cleaning up related methods. Update data response parsing to streamline functionality and improve clarity. Adjust imports accordingly. 2025-09-27 14:38:29 +03:00
vegu-ai-tools
1c7d28f83c Add gamestate context support in BuildPrompt and corresponding template. Introduced new property for gamestate context and updated rendering logic to include gamestate information in prompts. 2025-09-27 14:32:26 +03:00
vegu-ai-tools
5fb22467f2 prompt tweaks 2025-09-27 14:31:45 +03:00
vegu-ai-tools
b9f4d0a88a unified data extraction function 2025-09-27 14:31:17 +03:00
vegu-ai-tools
fb43310049 node updates 2025-09-27 11:43:05 +03:00
vegu-ai-tools
e98cafa63d Update message input hint in TalemateApp component to include keyboard shortcuts for navigating input history (Ctrl+Up/Down). 2025-09-26 22:24:54 +03:00
vegu-ai-tools
ae82e7ba2d Add input history functionality to message input in TalemateApp component. Implement keyboard shortcuts for navigating history (Ctrl+Up/Down) and limit history to the last 10 messages. Update message sending logic to store messages in history. 2025-09-26 22:24:06 +03:00
vegu-ai-tools
bad9453ba1 openroute fetch models on key set 2025-09-26 21:50:10 +03:00
vegu-ai-tools
f41506eeeb linting 2025-09-26 21:43:28 +03:00
vegu-ai-tools
6f47edbb27 Remove agent messages from state when opening agent message view in SceneTools component. 2025-09-26 21:43:13 +03:00
vegu-ai-tools
67facdf39e Add message emission for actor, narrator, and scene analysis guidance in respective components. Enhance AgentMessages and SceneTools for better message handling and visual feedback. 2025-09-26 21:39:37 +03:00
vegu-ai-tools
1fdc95e7cf Add agent state exclusions to changelog with a TODO for module migration 2025-09-26 21:39:23 +03:00
vegu-ai-tools
ab956f25a7 linting 2025-09-26 20:15:51 +03:00
vegu-ai-tools
cbbc843f86 Add AdvanceTime node to world state for time advancement with duration and narration instructions 2025-09-26 19:27:26 +03:00
vegu-ai-tools
028b0abf53 Update advance_time method to include return type annotation and return message 2025-09-26 19:27:13 +03:00
vegu-ai-tools
a7bbabbcad Add IsoDateDuration node for ISO 8601 interval string construction 2025-09-26 19:26:59 +03:00
vegu-ai-tools
e61baa6a5a narrate time action now has access to response length instructions 2025-09-26 19:26:50 +03:00
vegu-ai-tools
56b8e033ba Enhance type hints for duration conversion functions in time.py 2025-09-26 19:26:23 +03:00
vegu-ai-tools
1dd796aeb2 more prompt fixes 2025-09-26 18:23:35 +03:00
vegu-ai-tools
8d824acebc director chat prompt simplifications 2025-09-26 17:53:22 +03:00
vegu-ai-tools
2a1bd5864f prompt tweaks 2025-09-26 17:33:40 +03:00
vegu-ai-tools
e666cfe81a Update icons in NodeEditorLibrary and NodeEditorModuleProperties for improved UI clarity 2025-09-26 15:32:47 +03:00
vegu-ai-tools
d4182705e1 move module properties to navigation drawer 2025-09-26 15:30:59 +03:00
vegu-ai-tools
e57d5dbb73 linting 2025-09-26 14:12:23 +03:00
vegu-ai-tools
4fb5bc54ba move world state scene tools into sub component 2025-09-26 14:11:55 +03:00
vegu-ai-tools
f68610f315 remove leghacy world state manager buttons 2025-09-26 14:04:40 +03:00
vegu-ai-tools
e0e81a3796 Add cleanup function for recent scenes in config to remove non-existent paths 2025-09-26 14:00:39 +03:00
vegu-ai-tools
c53bbf2693 linting 2025-09-26 13:56:06 +03:00
vegu-ai-tools
9c60d4c046 restore from backup tweaks and scene loading error handling imrpvoements 2025-09-26 13:49:30 +03:00
vegu-ai-tools
a427b940b5 linting 2025-09-26 02:25:16 +03:00
vegu-ai-tools
63c992ab44 Refactor DirectorConsoleChatsToolbar to enhance UI with tooltips for persona and chat mode selection, improving user interaction and accessibility. 2025-09-26 02:24:54 +03:00
vegu-ai-tools
910764bde4 Add confirm write actions feature to chat context and UI components 2025-09-26 02:24:10 +03:00
vegu-ai-tools
7bbe5ced7e Implement auto-apply feature for input changes in GameState component, enhancing user experience by automatically committing changes after a brief delay. Update relevant methods to trigger auto-apply on various input events. 2025-09-26 02:13:45 +03:00
vegu-ai-tools
119dee7418 Enhance changelog functionality by adding delta type handling and improving commit behavior in InMemoryChangelog. Update tests to manually commit changes after appending deltas. 2025-09-26 02:04:29 +03:00
vegu-ai-tools
ddbb74a7a3 Implement unified refresh mechanism for active tabs in WorldStateManager, enhancing data loading for scene, characters, world, history, contextdb, pins, templates, and suggestions components. 2025-09-26 01:45:52 +03:00
vegu-ai-tools
64f7165fc8 refactor fork scene chip color and label in CharacterMessage and NarratorMessage components 2025-09-26 01:19:15 +03:00
vegu-ai-tools
5d008ae676 set default writing style in assistant mixin 2025-09-26 01:16:21 +03:00
vegu-ai-tools
62db9b1221 yield back to user on reject 2025-09-25 22:13:18 +03:00
vegu-ai-tools
a544367501 linting 2025-09-25 02:25:17 +03:00
vegu-ai-tools
314f24d23a pin decay 2025-09-25 02:25:10 +03:00
vegu-ai-tools
fbde0103bd linting 2025-09-25 01:43:26 +03:00
vegu-ai-tools
1c84ad76ea prompt tweaks 2025-09-25 01:41:24 +03:00
vegu-ai-tools
9c5d5cc322 set pins from context id 2025-09-25 01:41:04 +03:00
vegu-ai-tools
bf605604f0 append deltas on save 2025-09-25 01:38:51 +03:00
vegu-ai-tools
62d3aa25ca exclude world state from changelog 2025-09-25 01:38:39 +03:00
vegu-ai-tools
eb4e1426ac fixes 2025-09-24 22:42:09 +03:00
vegu-ai-tools
11f5242008 linting 2025-09-24 22:04:38 +03:00
vegu-ai-tools
4b0b252bfb Added scene activity check in DirectorChatActionConfirm to handle inactive scenes gracefully 2025-09-24 22:04:23 +03:00
vegu-ai-tools
85680a5285 linting 2025-09-24 21:24:01 +03:00
vegu-ai-tools
c8969d0fb7 fix UI flickering during quick agent workload swaps 2025-09-24 21:22:04 +03:00
vegu-ai-tools
3e5697d072 dleeting scene should remove changelog files 2025-09-24 21:13:52 +03:00
vegu-ai-tools
e446b01ac8 linting 2025-09-24 20:58:41 +03:00
vegu-ai-tools
1cc73f1899 memory changelog context 2025-09-24 20:58:33 +03:00
vegu-ai-tools
238911630b fork scene use changelog when available 2025-09-24 18:41:57 +03:00
vegu-ai-tools
89c9364db3 reset ltm after restore 2025-09-24 17:14:07 +03:00
vegu-ai-tools
b291afefd0 restore from backup ux polish 2025-09-24 16:10:29 +03:00
vegu-ai-tools
08850a7cb3 linting 2025-09-24 14:08:59 +03:00
vegu-ai-tools
9732e90a5b no need to init changelog during scene load 2025-09-24 14:08:30 +03:00
vegu-ai-tools
e689b18088 replace auto backup with restore from changelog 2025-09-24 14:04:47 +03:00
vegu-ai-tools
5123cbbef7 unix timestamps 2025-09-24 12:13:15 +03:00
vegu-ai-tools
b19e8cc645 store rev with scene messages 2025-09-24 12:06:03 +03:00
vegu-ai-tools
681102116d split changelog revisions 2025-09-24 11:36:35 +03:00
vegu-ai-tools
1b0f738e0b linting 2025-09-24 02:06:22 +03:00
vegu-ai-tools
eb094cc4b5 changelog integration progress 2025-09-24 02:01:15 +03:00
vegu-ai-tools
c434fb5a78 add deepdiff 2025-09-24 02:01:04 +03:00
vegu-ai-tools
6a37f673f6 changelog system 2025-09-24 02:00:53 +03:00
vegu-ai-tools
9cd5434a58 update argument_instructions and instructions fields to allow None values in Callback model 2025-09-23 23:22:16 +03:00
vegu-ai-tools
4fdebcb803 fix broken save world entry node 2025-09-23 17:12:34 +03:00
vegu-ai-tools
145be1096e bring manual commit back 2025-09-23 15:35:06 +03:00
vegu-ai-tools
05b3065ed2 auto commit 2025-09-23 15:22:52 +03:00
vegu-ai-tools
6d7b1cb063 ux tweaks 2025-09-23 13:31:59 +03:00
vegu-ai-tools
dd9a8f8ad4 linting 2025-09-23 02:25:59 +03:00
vegu-ai-tools
d4fcd724e3 cleanup 2025-09-23 02:25:49 +03:00
vegu-ai-tools
c0d3d7f14f gamestate editor 2025-09-23 01:58:38 +03:00
vegu-ai-tools
3ce834a432 linting 2025-09-22 13:47:23 +03:00
vegu-ai-tools
dc84c416b3 clean up rag config ux 2025-09-22 13:47:08 +03:00
vegu-ai-tools
e182a178c2 rag improvements 2025-09-22 13:33:20 +03:00
vegu-ai-tools
2b3251b46c always use semantic similarity retrieval since its fast 2025-09-22 12:57:52 +03:00
vegu-ai-tools
09f136dfb9 Add sentence compilation functions to dedupe module
- Introduced `compile_sentences_to_length` to join sentences into chunks of a specified length.
- Updated `__all__` to include new functions for improved usability in text processing.
2025-09-22 12:57:21 +03:00
vegu-ai-tools
3c7c5565f0 update confirm message 2025-09-22 00:09:18 +03:00
vegu-ai-tools
19af0e8156 Add read-only alert and conditional rendering in WorldStateManagerContextDB component
- Introduced a v-alert to notify users when the Context Database is in read-only mode.
- Updated button and table cell rendering to conditionally display based on the read-only state, enhancing user experience and clarity.
2025-09-22 00:07:06 +03:00
vegu-ai-tools
6baadc544d Remove the ability to add entries in WorldStateManagerContextDB component 2025-09-21 22:47:43 +03:00
vegu-ai-tools
bc072d9b68 fail safe for Listen nodes to avoid infinite failure cascade 2025-09-21 22:26:56 +03:00
vegu-ai-tools
67487b227e fix issue where inline function would not emit note state updates 2025-09-21 22:02:55 +03:00
vegu-ai-tools
4d4ca5e2cb ux tweaks 2025-09-21 21:57:14 +03:00
vegu-ai-tools
ed3f725b17 Add Copy Module Functionality to NodeEditor and NodeEditorLibrary
- Introduced a new button in NodeEditor for copying locked modules to the editable scene.
- Added a method to NodeEditorLibrary for pre-filling module details when copying to the scene, enhancing user workflow and module management.
2025-09-21 21:49:35 +03:00
vegu-ai-tools
4646bad50e Update Node Deletion Logic in NodeEditorLibrary
- Refined the condition for checking if the first module is deleted to ensure it only triggers when there are listed nodes, enhancing the robustness of node selection after deletions.
2025-09-21 20:50:29 +03:00
vegu-ai-tools
2fe1f4ff82 Enhance Node Deletion Handling in NodeEditorLibrary
- Updated logic to prevent selecting a deleted node as the first module in the list, ensuring a smoother user experience when nodes are removed.
2025-09-21 20:47:22 +03:00
vegu-ai-tools
533a618658 Remove console.log statements from DirectorConsoleChatMessageActionResult and NodeEditor components 2025-09-21 20:40:56 +03:00
vegu-ai-tools
d59807a68f - Introduced a unique ID for each Scene instance to enhance identification and tracking.
- Updated TalemateApp to manage the display of the new scene setup modal based on unique scene IDs, ensuring it only shows once per unique scene.
2025-09-21 01:21:00 +03:00
vegu-ai-tools
a497775de6 Enhance NodeEditor and TalemateApp Components
- Added 'toggle-scene-view' emit to NodeEditor for improved scene management.
- Wrapped SceneMessages component in a div within TalemateApp for better visibility control based on scene view state.
2025-09-21 01:12:06 +03:00
vegu-ai-tools
599d7115e0 Add New Scene Setup Modal Component
- Introduced a new modal for setting up scenes, allowing users to select writing styles and director personas.
- Integrated the modal into the TalemateApp component, enabling it to display when a new scene is detected.
- Added functionality to manage templates for writing styles and director personas within the modal.
- Implemented data handling for scene properties and user interactions, enhancing the user experience during scene creation.
2025-09-21 01:05:04 +03:00
veguAI
5f26134647 Director chat (#17)
* clarify instructions

* incrase default context length for attribute gen

* character progression node

* Call Agent Function conditional

* director chat tweaks

* fix issue where graph property editing would reload the graph and lose unsaved changes

* prompt tweaks

* character name optional

* use blur

* prompt tweaks

* director actions

* rename reason to instructions

* fix argument node type conversion

* prompt tweaks

* director action create character node improved

* nting

* scene budge and function name reorg

* memory nodes

* prompt tweaks

* get_arguments_node allow filter fn override

* query world information node

* smarter dict collector key values

* linting

* dedicated director action argument node

* node style

* FunctionWrapper find_nodes and first_node
CallForEach now suppoorts dict in items socket

* focal improvements

* world entry management nodes

* linting

* director action change world information

* instruction tweaks

* director action confirmation flow

* raise on reject

* director action confirmation progress

* single chat

* polish ux

* separation of components

* ux polish

* tweaks

* agent personas

* linting

* initial chat message override in persona

* agent persona to system message

* linting

* director chat compaction

* linting

* fix codeblock linebreaks

* prompt twaeks

* error message socket

* get scene types node

* collect None values

* linting

* director action nodes for scene direction management

* proimpt tweaks

* fix issue of director chat not working right on new scenes

* prompt tweaks

* director action summary node

* rename to content classification

* scene nodes to get/set intro, desc, title, content classification

* linting

* allow somew extra calls

* director action nodes

* fix query contextdb to use iterate value

* director action ndoes

* linting

* fix double cancellation issue on websocket plugin handlers

* fix node editor losing changes when switching to a different talemate tab and back

* fix resize handler

* fix group overlap bug during snap action

* clear validation messages from property editor

* improve node search matching

* fix dynamic socket issues

* cleanup

* allow hot reload of new DA or command modules

* linting

* fix tests

* director modes

* allow changing director persona from chat interface

* tweaks

* separate state reinf component

* cleanup

* separate components for char attrib and details

* separate component for spirce collection

* separate writing style component

* cleanup

* remove applicable_agents

* persist chat mode

* linting

* ux tweaks

* director chat better context managemnet

* wording

* displat budgets in UX

* reorg

* Validate Is Not Set Node

* character nodes

* add extra output sockets

* fix compact error

* fix compact error

* fancy diffs

* nodes updated

* summarizer settings node

* fix type hint

* add useful output sockets

* history archive nodes

* add dedupe_enabled and response_length properties

* prompt tweaks

* linting

* nodes

* prompt tweaks

* prompt tweaks

* linting

* better instruct_character actions

* fix Get node to work with tuples and sets

* query character information should include details

* lint

* tweak instructions

* context id impl

* context id

* fix registry id

* context id socket

* context id

* build prompt improvements

* extract list node

* context_id as socket type

* remove empty separators

* linting

* character cibtext

* Fix advanced format always executing

* expose context id on character attrib

* CombineList node

* Dynamic instructions ndoe can now be fed list of strings

* return the context id object

* expose context id when unpacking memory doc

* progress on improved direction query action

* linting

* tweaks

* fix dynamic sockets not being copied during node clone

* fix dynamic sockets being lost when creating module from selection

* fix nodes spawning in too small to contain title

* sort choices

* hide prop value if related socket is connected

* shorten character context ids

* fix ai function type conversion issue that would cast everything to str

* hash context id

* context id value

* tests

* linting

* rename and tests

* linting

* refactor context id handler a bit

* context id shenanigans

* fix tests

* cleanup

* remove unused imports

* history context id handler

* refactor context id into proper module structure

* linting

* retrieve context

* world entry context ids

* story config context ids

* story config context

* linting

* no longer needed

* context id progress

* fix tests

* scene type inspection context ids

* linting

* prompt tweaks

* prompt tweaks

* shift+alt drag node for counterpart creation

* node property editor will now confirm close if it has changes

* transfer name property

* node coutnerpart fixes

* counterpart copy size fixes

* character_status socket

* fix director confirm node error when called outside of director chat context

* if input and output socket counterpart def

* prompt tweaks

* director action nodes

* no longer needed

* instruct character creation

* fix title

* toggle character

* linting

* GPT-OSS base templte

* wass reasoning_tokens to model_prompt

* pass reasoning_tokens to model prompt template

* gpt-oss preset

* to warning

* prompt tweaks

* prompt tweaks

* wording

* prompt tweaks

* pass through error message

* new exceptions

* clean up

* status field

* better response parsing

* linting

* add sockets and field to GetWorldEntry

* auto size node to fit all the widgets when adding to the graph

* contextual generate node correctly validates context_name as required

* fix issue where alt dragging a single node wouldnt work if other nodes were selected

* create group from node selection clears the selection after creation

* group from selected nodes - presets

* fix ctrl enter in text properties adding extra newline

* mark unresolved required sockets

* linting

* fix issue where connections where incorrectly flagged as unresolved

* Add GetCharacterDetail and SetCharacterDetail nodes to manage character details

- Introduced GetCharacterDetail node to retrieve character details based on a name input.
- Added SetCharacterDetail node to set character details with specified name and value.
- Updated existing GetCharacterAttribute node to handle cases where attributes may not exist, ensuring safe access to context IDs.

* Add Character Context Module and Update Instruct Character Updates

- Introduced a new character context module with various nodes for managing character attributes, details, and descriptions.
- Removed obsolete nodes from the instruct-character-updates module to streamline functionality.
- Adjusted node positions and properties for better organization and clarity in the graph structure.

* linting

* some context id stuff

* linting

* determine character dialogue instructions can now rewrite existing instructions

* allow none values

* context id node fixes

* prompt tweaks

* Add CounterPart functionality to DefineFunction and GetFunction nodes

* linting

* character config updates

* module sync

* dialogue_instructions -> acting_instructions

* linting

* story conig updates

* fix tests

* remove old action

* scan context ids

* director action tweaks

* Add scene_type_ids output to GetSceneTypes node

* director action nodes

* linting

* director agent nodes

* director action direct scene

* nodes

* nodes

* context id collector

* linting

* Handle empty content in DynamicInstruction string representation

* Add new color items "Prepare" and "Special" to recent nodes context menu

* Rename and separate hook processing methods in FocalContext for clarity

* Refactor action result handling in DirectorChatMixin to improve feedback mechanism and streamline chat state updates

* Add custom instructions feature to DirectorChatMixin and update chat instructions template to display them. Refactor existing action configurations for clarity.

* Add chat common tasks and limitations template for Director. Include scenarios for story creation, character behavior, memory issues, and repetition handling.

* Update chat template to clarify the role of the Director and include Talemate system context. Add common tasks template for enhanced chat functionality.

* prompt tweaks

* Add scene code block processing to DirectorConsoleChatMessageMarkdown component. Enhance markdown rendering by integrating scene text parsing and updating styles for scene blocks.

* Enhance NodeEditor and NodeEditorLibrary components with a new module library drawer and improved node display. Introduce tree view for better organization of modules, including scenes and agents, and update node display labels for clarity. Refactor resizing logic for the editor container.

* Implement scene view toggle and exit confirmation in NodeEditor. Move creative mode controls from TalemateApp to NodeEditor, enhancing user experience with new buttons and confirmation prompts for exiting creative mode.

* linting
2025-09-21 00:18:57 +03:00
vegu-ai-tools
b971c3044d restore from backup function 2025-08-30 01:06:20 +03:00
vegu-ai-tools
f3d02530d5 linting 2025-08-30 00:39:13 +03:00
vegu-ai-tools
9bf08b1f00 auto backup 2025-08-30 00:38:18 +03:00
vegu-ai-tools
b0f1b7307c reserved property names 2025-08-30 00:02:34 +03:00
vegu-ai-tools
9134c0cc26 node styles 2025-08-26 22:00:08 +03:00
vegu-ai-tools
3141e53eac dont need this 2025-08-26 21:52:11 +03:00
vegu-ai-tools
307439b210 tweaks and tests 2025-08-26 21:50:05 +03:00
vegu-ai-tools
ab3f4f3b2e dict collector improvements 2025-08-26 21:36:15 +03:00
vegu-ai-tools
b862159aef streamline add / remove dyn socket 2025-08-26 21:18:20 +03:00
vegu-ai-tools
8f4aa75e09 fixes 2025-08-26 21:06:26 +03:00
vegu-ai-tools
7320196ac6 garbage 2025-08-26 16:44:35 +03:00
vegu-ai-tools
e7b949c443 set 0.33.0 2025-08-26 10:20:35 +03:00
vegu-ai-tools
7ffbfe8d0d fix alt-drag to clone single node 2025-08-26 10:15:43 +03:00
348 changed files with 5761 additions and 34579 deletions

View File

@@ -19,25 +19,6 @@ jobs:
steps:
- uses: actions/checkout@v4
- name: Remove unnecessary files to release disk space
run: |
sudo rm -rf \
"$AGENT_TOOLSDIRECTORY" \
/opt/ghc \
/opt/google/chrome \
/opt/microsoft/msedge \
/opt/microsoft/powershell \
/opt/pipx \
/usr/lib/mono \
/usr/local/julia* \
/usr/local/lib/android \
/usr/local/lib/node_modules \
/usr/local/share/chromium \
/usr/local/share/powershell \
/usr/local/share/powershell \
/usr/share/dotnet \
/usr/share/swift
- name: Log in to GHCR
uses: docker/login-action@v3
with:

View File

@@ -14,25 +14,6 @@ jobs:
steps:
- uses: actions/checkout@v4
- name: Remove unnecessary files to release disk space
run: |
sudo rm -rf \
"$AGENT_TOOLSDIRECTORY" \
/opt/ghc \
/opt/google/chrome \
/opt/microsoft/msedge \
/opt/microsoft/powershell \
/opt/pipx \
/usr/lib/mono \
/usr/local/julia* \
/usr/local/lib/android \
/usr/local/lib/node_modules \
/usr/local/share/chromium \
/usr/local/share/powershell \
/usr/local/share/powershell \
/usr/share/dotnet \
/usr/share/swift
- name: Log in to GHCR
uses: docker/login-action@v3
with:

View File

@@ -45,9 +45,6 @@ WORKDIR /app
RUN apt-get update && apt-get install -y \
bash \
wget \
tar \
xz-utils \
&& rm -rf /var/lib/apt/lists/*
# Install uv in the final stage
@@ -56,21 +53,6 @@ RUN pip install uv
# Copy virtual environment from backend-build stage
COPY --from=backend-build /app/.venv /app/.venv
# Download and install FFmpeg 8.0 with shared libraries into .venv (matching Windows installer approach)
# Using BtbN FFmpeg builds which provide shared libraries - verified to work
# Note: We tried using jrottenberg/ffmpeg:8.0-ubuntu image but copying libraries from it didn't work properly,
# so we use the direct download approach which is more reliable and matches the Windows installer
RUN cd /tmp && \
wget -q https://github.com/BtbN/FFmpeg-Builds/releases/download/latest/ffmpeg-master-latest-linux64-gpl-shared.tar.xz -O ffmpeg.tar.xz && \
tar -xf ffmpeg.tar.xz && \
cp -a ffmpeg-master-latest-linux64-gpl-shared/bin/* /app/.venv/bin/ && \
cp -a ffmpeg-master-latest-linux64-gpl-shared/lib/* /app/.venv/lib/ && \
rm -rf ffmpeg-master-latest-linux64-gpl-shared ffmpeg.tar.xz && \
LD_LIBRARY_PATH=/app/.venv/lib /app/.venv/bin/ffmpeg -version | head -n 1
# Set LD_LIBRARY_PATH so torchcodec can find ffmpeg libraries at runtime
ENV LD_LIBRARY_PATH=/app/.venv/lib:${LD_LIBRARY_PATH}
# Copy Python source code
COPY --from=backend-build /app/src /app/src

View File

@@ -0,0 +1,134 @@
"""
An attempt to write a client against the runpod serverless vllm worker.
This is close to functional, but since runpod serverless gpu availability is currently terrible, i have
been unable to properly test it.
Putting it here for now since i think it makes a decent example of how to write a client against a new service.
"""
import pydantic
import structlog
import runpod
import asyncio
import aiohttp
from talemate.client.base import ClientBase, ExtraField
from talemate.client.registry import register
from talemate.emit import emit
from talemate.config import Client as BaseClientConfig
log = structlog.get_logger("talemate.client.runpod_vllm")
class Defaults(pydantic.BaseModel):
max_token_length: int = 4096
model: str = ""
runpod_id: str = ""
class ClientConfig(BaseClientConfig):
runpod_id: str = ""
@register()
class RunPodVLLMClient(ClientBase):
client_type = "runpod_vllm"
conversation_retries = 5
config_cls = ClientConfig
class Meta(ClientBase.Meta):
title: str = "Runpod VLLM"
name_prefix: str = "Runpod VLLM"
enable_api_auth: bool = True
manual_model: bool = True
defaults: Defaults = Defaults()
extra_fields: dict[str, ExtraField] = {
"runpod_id": ExtraField(
name="runpod_id",
type="text",
label="Runpod ID",
required=True,
description="The Runpod ID to connect to.",
)
}
def __init__(self, model=None, runpod_id=None, **kwargs):
self.model_name = model
self.runpod_id = runpod_id
super().__init__(**kwargs)
@property
def experimental(self):
return False
def set_client(self, **kwargs):
log.debug("set_client", kwargs=kwargs, runpod_id=self.runpod_id)
self.runpod_id = kwargs.get("runpod_id", self.runpod_id)
def tune_prompt_parameters(self, parameters: dict, kind: str):
super().tune_prompt_parameters(parameters, kind)
keys = list(parameters.keys())
valid_keys = ["temperature", "top_p", "max_tokens"]
for key in keys:
if key not in valid_keys:
del parameters[key]
async def get_model_name(self):
return self.model_name
async def generate(self, prompt: str, parameters: dict, kind: str):
"""
Generates text from the given prompt and parameters.
"""
prompt = prompt.strip()
self.log.debug("generate", prompt=prompt[:128] + " ...", parameters=parameters)
try:
async with aiohttp.ClientSession() as session:
endpoint = runpod.AsyncioEndpoint(self.runpod_id, session)
run_request = await endpoint.run(
{
"input": {
"prompt": prompt,
}
# "parameters": parameters
}
)
while (await run_request.status()) not in [
"COMPLETED",
"FAILED",
"CANCELLED",
]:
status = await run_request.status()
log.debug("generate", status=status)
await asyncio.sleep(0.1)
status = await run_request.status()
log.debug("generate", status=status)
response = await run_request.output()
log.debug("generate", response=response)
return response["choices"][0]["tokens"][0]
except Exception as e:
self.log.error("generate error", e=e)
emit(
"status", message="Error during generation (check logs)", status="error"
)
return ""
def reconfigure(self, **kwargs):
if kwargs.get("model"):
self.model_name = kwargs["model"]
if "runpod_id" in kwargs:
self.api_auth = kwargs["runpod_id"]
self.set_client(**kwargs)

Binary file not shown.

Before

Width:  |  Height:  |  Size: 346 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 702 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 5.5 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 54 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 9.4 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.0 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 42 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 19 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 18 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 471 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 180 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 82 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 48 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 30 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 70 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 46 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 45 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 4.9 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 24 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 411 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 18 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 9.5 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 21 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 16 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 27 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 28 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 371 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 24 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 46 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 45 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 2.8 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 50 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 42 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 44 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 41 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.9 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.0 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 44 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 35 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 50 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 30 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 25 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 22 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 4.4 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 50 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 27 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 25 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 21 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 4.7 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 52 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 48 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 47 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 45 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 4.5 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 48 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 51 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 6.6 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 54 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.9 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 6.9 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 20 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 20 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 45 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 53 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 60 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 11 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.1 MiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 15 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 29 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.0 MiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 24 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 15 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 5.9 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 91 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 39 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 2.5 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 24 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 23 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.9 MiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 16 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 22 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.0 MiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 32 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 8.0 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 4.6 KiB

View File

@@ -0,0 +1,42 @@
# AUTOMATIC1111
!!! info
This requires you to setup a local instance of the AUTOMATIC1111 API. Follow the instructions from [their GitHub](https://github.com/AUTOMATIC1111/stable-diffusion-webui) to get it running.
Once you have it running, you will want to adjust the `webui-user.bat` in the AUTOMATIC1111 directory to include the following command arguments:
```bat
set COMMANDLINE_ARGS=--api --listen --port 7861
```
Then run the `webui-user.bat` to start the API.
Once your AUTOAMTIC1111 API is running (check with your browser) you can set the Visualizer config to use the `AUTOMATIC1111` backend
## Settings
![Visual agent automatic1111 settings](/talemate/img/0.27.0/automatic1111-settings.png)
##### API URL
The url of the API, if following this example, should be `http://localhost:7861`
##### Steps
The number of steps to use for image generation. More steps will result in higher quality images but will take longer to generate.
##### Sampling Method
Which sampling method to use for image generation.
##### Schedule Type
Which scheduler to use for image generation.
##### CFG Scale
CFG scale for image generation.
##### Model type
Differentiates between `SD1.5` and `SDXL` models. This will dictate the resolution of the image generation and actually matters for the quality so make sure this is set to the correct model type for the model you are using.

View File

@@ -1,80 +0,0 @@
# AUTOMATIC1111
!!! warning "Deprecated Backend"
**AUTOMATIC1111 (A1111) is essentially dead at this point** - development has largely stopped and the project is no longer actively maintained. Support for AUTOMATIC1111 has only been carried forward in Talemate because it was easy to maintain compatibility.
**We strongly recommend using [SD.Next](sdnext.md) instead**, which is an actively maintained fork of AUTOMATIC1111 with improved performance, better features, and ongoing development. SD.Next maintains API compatibility with AUTOMATIC1111, so migration is straightforward.
The AUTOMATIC1111 backend provides basic text-to-image generation capabilities using the AUTOMATIC1111 Stable Diffusion WebUI API. This backend only supports text-to-image generation - it does not support image editing or image analysis.
![This screenshot displays the dark-themed "General" settings interface for an application named "Visualizer," featuring a sidebar menu on the left and configuration options on the right. Key settings include dropdowns where the Client is set to "Google" and the text-to-image backend is set to "AUTOMATIC1111," alongside an image generation timeout slider positioned at 301. Additional controls show a checked box for "Automatic Setup," an unchecked box for "Automatic Generation," and a "Fallback Prompt Type" set to "Keywords."](/talemate/img/0.34.0/visual-agent-a1111-1.png)
## Prerequisites
Before configuring the AUTOMATIC1111 backend, you need to have AUTOMATIC1111 installed and running:
1. Install and start AUTOMATIC1111 Stable Diffusion WebUI on your system
2. Ensure the API is enabled and accessible
3. Note the API URL (default is `http://localhost:7860`)
!!! note "Migration to SD.Next"
If you're setting up a new installation, please use [SD.Next](sdnext.md) instead. If you have an existing AUTOMATIC1111 installation, consider migrating to SD.Next for better performance and ongoing support.
## Configuration
In the Visualizer agent settings, select AUTOMATIC1111 as your backend for text-to-image generation.
### Text-to-Image Configuration
For text-to-image generation, configure the following settings:
- **API URL**: The URL where your AUTOMATIC1111 instance is running (e.g., `http://localhost:7860`)
- **Steps**: Number of sampling steps (default: 40, range: 5-150)
- **Sampling Method**: The sampling algorithm to use (e.g., "DPM++ 2M", "Euler a")
- **Schedule Type**: The sampling schedule to use (e.g., "Automatic", "Karras", "Uniform")
- **CFG Scale**: Classifier-free guidance scale (default: 7.0, range: 1-30)
- **Prompt Type**: Choose between "Keywords" or "Descriptive" prompt formatting
- **Resolutions**: Configure the pixel dimensions for Square, Portrait, and Landscape formats
![A screenshot of the Visualizer interface displaying configuration settings for Automatic1111 text-to-image generation. The panel features adjustable parameters such as the API URL, sampler steps, CFG scale, and sampling method. Additionally, it includes sections for selecting prompting types and setting default resolutions for square, portrait, and landscape image orientations.](/talemate/img/0.34.0/visual-agent-a1111-2.png)
!!! note "No Authentication"
AUTOMATIC1111 backend does not support authentication. If your AUTOMATIC1111 instance requires authentication, you'll need to either disable it or use SD.Next instead, which supports authentication.
!!! note "Model Selection"
AUTOMATIC1111 does not support model selection through the API. The backend will use whatever model is currently loaded in your AUTOMATIC1111 instance. You need to change models manually in the AUTOMATIC1111 WebUI interface.
## Usage
Once configured, the AUTOMATIC1111 backend will appear in the Visualizer agent status with a green indicator showing text-to-image capability is available.
![A dark mode interface element titled "Visualizer" accompanied by a green status dot. Below the title are two badges: a gray button labeled "Google" with a computer icon and a green button labeled "AUTOMATIC1111" with an image icon.](/talemate/img/0.34.0/visual-agent-a1111-3.png)
## Limitations
The AUTOMATIC1111 backend has several limitations compared to SD.Next:
- **No image editing**: Only supports text-to-image generation
- **No authentication**: Cannot connect to instances that require authentication
- **No model selection**: Uses whatever model is loaded in AUTOMATIC1111
- **No active development**: The AUTOMATIC1111 project is no longer actively maintained
## Sampler Settings
AUTOMATIC1111 provides control over the generation process:
- **Steps**: More steps generally produce higher quality images but take longer. Typical values range from 20-50 steps, with 40 being a good default.
- **Sampling Method**: Different samplers produce different results. Popular options include:
- **DPM++ 2M**: Fast and high quality (default)
- **Euler a**: Fast, good for quick iterations
- **DPM++ SDE**: Variant with different characteristics
- **Schedule Type**: Controls the noise schedule used during sampling. "Automatic" is typically the best choice.
- **CFG Scale**: Controls how closely the model follows your prompt. Lower values (1-7) allow more creative freedom, higher values (7-15) stick closer to the prompt.
## Prompt Formatting
AUTOMATIC1111 uses **Keywords** prompt formatting by default. This means prompts are formatted as keyword lists optimized for Stable Diffusion models. You can switch to **Descriptive** formatting if you prefer natural language descriptions, though Keywords typically work better with SD models.
## Automatic Setup with KoboldCpp
If you're using KoboldCpp with AUTOMATIC1111 support, Talemate can automatically detect and configure the AUTOMATIC1111 backend when "Automatic Setup" is enabled in the Visualizer settings. This will automatically set the API URL to match your KoboldCpp instance URL.

View File

@@ -1,166 +0,0 @@
# ComfyUI
## Prepare ComfyUI
This document assumes you have installed ComfyUI (either the portable or the desktop version).
Copy the .bat file you use to start ComfyUI and add the `--port` parameter.
```
--port 8188
```
You can put any port you want, but this example will use 8188.
!!! note "If you are using a remote ComfyUI instance"
If you are using a remote ComfyUI instance, you may want to add the `--listen` parameter as well.
```
--listen 0.0.0.0
```
You will then also need to obtain the IP address of the computer running ComfyUI and use it in the Talemate configuration. (instead of localhost)
Confirm ComfyUI is running in your browser by visiting http://localhost:8188 or `http://<ip-address>:8188` before proceeding to talemate.
## Talemate configuration
In the Visualizer agent settings, select ComfyUI as your backend for text-to-image generation, image editing, or both. You'll need to configure each backend separately if you want to use ComfyUI for different operations.
![The image displays the General settings tab of the Visualizer interface, featuring a sidebar with active indicators for ComfyUI and Google modules. Dropdown menus in the main panel show ComfyUI selected for text-to-image and image editing backends, with Google selected for image analysis. The interface also includes an image generation timeout slider set to 301 and an enabled Automatic Setup checkbox.](/talemate/img/0.34.0/visual-agent-comfyui-1.png)
### Text-to-Image Configuration
For text-to-image generation, configure the following settings:
- **API URL**: The URL where your ComfyUI instance is running (e.g., `http://localhost:8188`)
- **Workflow**: Select the workflow file to use for generation. Talemate includes several pre-configured workflows including `qwen_image.json` and `z_image_turbo.json`
- **Model**: Select the model to use from your ComfyUI models directory. If your workflow doesn't include a "Talemate Load Model" or "Talemate Load Checkpoint" node, this will be set to "- Workflow default -" and the model specified in the workflow file will be used.
- **Prompt Type**: Choose between "Keywords" or "Descriptive" prompt formatting
!!! tip "Choosing Prompt Type"
As a general rule: **SDXL models** typically work best with **Keywords** formatting, while most other models (including Qwen Image, Flux, etc.) work better with **Descriptive** formatting. If you're unsure, start with Descriptive and switch to Keywords if you're using an SDXL-based workflow.
- **Resolutions**: Configure the pixel dimensions for Square, Portrait, and Landscape formats
![A screenshot of the "Visualizer" application's dark-mode settings panel specifically for ComfyUI text-to-image generation. The interface features configuration fields for the API URL, a workflow dropdown set to "z_image_turbo.json," model selection, and a "Descriptive" prompting type. The lower section includes adjustable numeric inputs for defining pixel dimensions for Square, Portrait, and Landscape image resolutions.](/talemate/img/0.34.0/visual-agent-comfyui-2.png)
![This screenshot displays the dark-themed settings interface of an application named "Visualizer," specifically configured for ComfyUI text-to-image generation. The main panel features input fields for the API URL, workflow selection (set to default-sdxl), and model choice (juggernautXL), along with a prompting type setting. Below these options is a "Resolutions" section allowing users to define specific pixel dimensions for Square, Portrait, and Landscape image outputs.](/talemate/img/0.34.0/visual-agent-comfyui-3.png)
### Image Editing Configuration
For image editing, configure similar settings but select an image editing workflow such as `qwen_image_edit.json`. The number of reference images supported depends on your model - for example, Qwen Image Edit can handle up to 3 reference images that can be used to guide the editing process.
!!! note "Prompt Type for Image Editing"
Image editing workflows typically use **Descriptive** prompt formatting by default, as most image editing models (like Qwen Image Edit) work better with descriptive instructions rather than keyword-based prompts.
![A screenshot of the "Visualizer" application settings interface, specifically showing the configuration panel for "ComfyUI Image Editing." The main view displays input fields for the API URL, a selected workflow file named "qwen_image_edit.json," descriptive prompting settings, and resolution presets for square, portrait, and landscape aspect ratios.](/talemate/img/0.34.0/visual-agent-comfyui-4.png)
![This screenshot shows a browser tab group labeled "Visualizer" marked with a green status dot on a dark background. The group contains four tabs: a Google link, two green-tinted ComfyUI tabs with image and pencil icons, and a gray tab titled "References 3".](/talemate/img/0.34.0/visual-agent-comfyui-5.png)
## Custom workflow creation
Talemate comes with pre-configured workflows for Qwen Image models (`qwen_image.json` for text-to-image and `qwen_image_edit.json` for image editing). However, since there are many variables in ComfyUI setups (different model formats like GGUF vs safetensors, custom LoRAs, different hardware configurations, etc.), you may want to customize these workflows to match your specific setup.
### Starting from a Template
Open ComfyUI in your browser and navigate to the templates menu. ComfyUI includes workflow templates that you can use as a starting point:
- **Qwen Image**: For text-to-image generation
- **Qwen Image Edit**: For image editing workflows
These templates provide a good foundation for creating custom workflows.
![A dark-themed dropdown menu from a software interface is shown, featuring a header labeled "image_qwen_image." The menu lists standard options such as New, File, Edit, View, and Theme, followed by specific actions like Browse Templates, Settings, Manage Extensions, and Help.](/talemate/img/0.34.0/comfyui.workflow.setup.browse-templates.png)
![A product card for the "Qwen-Image Text to Image" AI model, displaying a sample generation of a rainy, neon-lit street scene with vibrant pink and blue signage. The image demonstrates the model's capabilities by clearly rendering complex multilingual text, such as Chinese characters and English words like "HAPPY HAIR," on the storefronts. Below the visual, a brief description highlights the tool's exceptional text rendering and editing features.](/talemate/img/0.34.0/comfyui.workflow.setup.qwen-template.png)
Load the Qwen Image template to see the base workflow structure.
![A screenshot of a ComfyUI workflow designed for the Qwen-Image diffusion model, featuring grouped nodes for model loading, image sizing, and text prompting. The interface includes detailed instructional notes regarding VRAM usage on an RTX 4090D, model storage locations, and optimal KSampler settings. A positive prompt node is visible containing a detailed description of a neon-lit Hong Kong street scene.](/talemate/img/0.34.0/comfyui.workflow.setup.qwen-start.png)
### Naming Nodes for Talemate
For Talemate to properly interact with your workflow, you need to rename specific nodes with exact titles. These titles allow Talemate to inject prompts, set resolutions, and handle reference images automatically.
**Required Node Titles:**
1. **Talemate Positive Prompt**: The node that encodes the positive prompt (typically a `CLIPTextEncode` or `TextEncodeQwenImageEditPlus` node). This is required - workflows without this node will fail validation.
2. **Talemate Negative Prompt**: The node that encodes the negative prompt (same node types as above)
3. **Talemate Resolution**: The node that sets the image dimensions (typically an `EmptySD3LatentImage` or similar latent image node)
**Optional Node Titles:**
- **Talemate Load Model** or **Talemate Load Checkpoint**: If you want to allow model selection from Talemate's settings, rename your model loader node (typically `CheckpointLoaderSimple`, `UNETLoader`, or `UnetLoaderGGUF`) to one of these titles. If this node is not present, Talemate will use the model specified in the workflow file itself, and the model dropdown will show "- Workflow default -" as the only option.
To rename a node, right-click on it and select "Rename" or double-click the node title, then enter the exact title name.
![A screenshot of a node-based interface labeled "Step 3 - Prompt," featuring a green "Talemate Positive Prompt" node containing a detailed text description of a vibrant, neon-lit Hong Kong street scene. The text specifies a 1980s cinematic atmosphere and lists numerous specific shop signs in both Chinese and English. Below it, a dark red "Talemate Negative Prompt" node is visible but currently contains no text.](/talemate/img/0.34.0/comfyui.workflow.setup.talemate-prompts.png)
![This image displays a dark green interface node labeled "Talemate Positive Prompt," typical of a node-based editor like ComfyUI. It features a yellow input connection point for "clip" on the left, an orange output point for "CONDITIONING" on the right, and a large, dark text entry field in the center containing the placeholder word "text".](/talemate/img/0.34.0/comfyui.workflow.setup.talemate-empty-prompt.png)
![A screenshot of a dark gray interface node labeled "Talemate Resolution" with the identifier #58. It features configurable fields for width and height, both set to 1328, and a batch size of 1. The node has a single output connection point labeled "LATENT".](/talemate/img/0.34.0/comfyui.workflow.setup.talemate-resulotion.png)
### Activating the Lightning LoRA (Optional)
The Qwen Image template includes a Lightning LoRA node that is deactivated by default. You can optionally activate it to speed up generation with fewer steps. Note that this is a trade-off: the Lightning LoRA reduces generation time but may degrade image quality compared to using more steps without the LoRA.
To activate the Lightning LoRA:
1. Find the `LoraLoaderModelOnly` node in your workflow (it should already be present in the Qwen template)
2. Connect it between your model loader and sampler if it's not already connected
3. Load the appropriate Lightning LoRA file (e.g., `Qwen-Image-Lightning-8steps-V1.0.safetensors` for 8-step generation)
4. Adjust your sampler settings:
- **Steps**: Reduce to 8 steps (or 4 steps for the 4-step variant)
- **CFG Scale**: Set to 1.0 (lower than typical values)
![This screenshot features a "LoraLoaderModelOnly" node within a ComfyUI workflow, customized with the label "Lightx2v 8steps LoRA". It shows the selection of a "Qwen-Image-Lightning-8steps" LoRA file with a model strength parameter set to 1.00. Purple connection cables are visible attached to the input and output model terminals.](/talemate/img/0.34.0/comfyui.workflow.setup.lighting-lora.png)
![The image shows a close-up of a dark user interface panel containing two adjustable setting fields. The top field is labeled "steps" and displays a value of 8, flanked by left and right directional arrows. Below that, a second field labeled "cfg" shows a value of 1.0, also with adjustment arrows on either side.](/talemate/img/0.34.0/comfyui.workflow.setup.lighting-lora-sampler-changes.png)
### Image Editing Workflows: Reference Nodes
For image editing workflows (like `qwen_image_edit.json`), you need to add reference image nodes. Note that ComfyUI includes a Qwen Image Edit template similar to the Qwen Image template, which you can use as a starting point.
!!! warning "Reference Nodes Required"
Image editing workflows **must** define at least one reference node. If your workflow doesn't include any nodes titled "Talemate Reference 1" (or higher), the backend status will show an error and image editing will not work.
These are `LoadImage` nodes that Talemate will use to inject reference images for editing.
The number of reference nodes you can add depends on your model's capabilities. For example, Qwen Image Edit supports up to 3 reference images. Add `LoadImage` nodes and rename them with these exact titles:
- **Talemate Reference 1**
- **Talemate Reference 2**
- **Talemate Reference 3** (if your model supports it)
These nodes should be connected to your prompt encoding nodes (for Qwen Image Edit, use `TextEncodeQwenImageEditPlus` nodes that accept image inputs).
![Three identical interface nodes labeled "Talemate Reference 1," "2," and "3" are arranged horizontally within a dark-themed node-based editor. Each node features output ports for "IMAGE" and "MASK," along with a file selection field showing "image_qwen_image_edit" and a "choose file to upload" button. Blue and red connection wires link these nodes to other off-screen elements in the workflow.](/talemate/img/0.34.0/comfyui.workflow.setup.talemate-references.png)
### Saving and Exporting the Workflow
Once your workflow is configured, you need to save it and export it in the API format for Talemate to use it.
1. **Save the workflow**: Use File → Save As to save your workflow as a `.json` file in your ComfyUI workflows directory
2. **Export for API**: Use File → Export (API) to create the API-compatible version
!!! warning "Export vs Export (API)"
It's critical to use **"Export (API)"** and not just "Export". The regular export format is not compatible with Talemate's API integration. The API export format includes the necessary metadata and structure that Talemate expects.
![A screenshot of a dark-themed software interface menu with the "File" option selected, revealing a nested sub-menu. The sub-menu lists file management commands, with the "Save As" option highlighted among choices like Open, Save, and Export.](/talemate/img/0.34.0/comfyui.workflow.setup.qwen-save.png)
![This image displays a dark-themed user interface menu, likely from ComfyUI, with the "File" category expanded. A submenu lists options including Open, Save, and Save As, while the "Export (API)" option is currently highlighted at the bottom. This visual illustrates how to locate the API export function within the software's file management system.](/talemate/img/0.34.0/comfyui.workflow.setup.qwen-export.png)
After exporting, place the workflow JSON file in Talemate's `templates/comfyui-workflows` directory. Once placed there, it will automatically appear in the workflow dropdown in Talemate's ComfyUI settings.
!!! note "Workflow File Location"
Workflow files must be placed in Talemate's `templates/comfyui-workflows` directory, not ComfyUI's workflows directory. Talemate loads workflows from its own templates directory to ensure compatibility and proper integration.
!!! tip "Workflow Not Appearing?"
If your workflow file doesn't appear in the agent's settings dropdown after placing it in the correct directory, try reloading the Talemate browser window. The workflow list is refreshed when the page loads.
!!! info "Hot-Reloading Workflows"
Changes to workflow files are automatically detected and reloaded by the agent. After modifying a workflow file, your changes will be applied to the next image generation without needing to restart Talemate or reload the browser window.

View File

@@ -1,101 +0,0 @@
# Google
The Google backend provides image generation, editing, and analysis capabilities using Google's Gemini image models. It supports text-to-image generation, image editing with reference images, and AI-powered image analysis.
![A screenshot of the "Visualizer" application settings interface with the "General" tab selected. It shows configuration dropdowns for Client and various Backends (text to image, image editing, image analysis) all set to "Google," alongside an image generation timeout slider positioned at 301. Additional settings include a checked "Automatic Setup" box, an unchecked "Automatic Generation" box, and a "Fallback Prompt Type" menu set to "Keywords."](/talemate/img/0.34.0/visual-agent-google-4.png)
## Prerequisites
Before configuring the Google backend, you need to obtain a Google API key:
1. Go to [Google AI Studio](https://aistudio.google.com/app/apikey)
2. Sign in with your Google account
3. Create a new API key or use an existing one
4. Copy the API key
Then configure it in Talemate:
1. Open Talemate Settings → Application → Google
2. Paste your Google API key in the "Google API Key" field
3. Save your changes
!!! note "API Key vs Vertex AI Credentials"
The Visualizer agent uses the Google API key (not Vertex AI service account credentials). Make sure you're using the API key from Google AI Studio, not the service account JSON file used for Vertex AI.
## Configuration
In the Visualizer agent settings, select Google as your backend for text-to-image generation, image editing, image analysis, or any combination of these. Each operation can be configured separately.
### Text-to-Image Configuration
For text-to-image generation, configure the following settings:
- **Google API Key**: Your Google API key (configured globally in Talemate Settings)
- **Model**: Select the image generation model to use:
- **gemini-2.5-flash-image**: Faster generation, good quality
- **gemini-3-pro-image-preview**: Higher quality, slower generation
![A dark-themed settings interface for a "Visualizer" application, specifically showing the "Google Text to Image" configuration panel. The main view displays a masked input field for a configured Google API Key and a dropdown menu selecting the "gemini-3-pro-image-preview" model.](/talemate/img/0.34.0/visual-agent-google-5.png)
The Google backend automatically handles aspect ratios based on the format you select:
- **Landscape**: 16:9 aspect ratio
- **Portrait**: 9:16 aspect ratio
- **Square**: 1:1 aspect ratio
### Image Editing Configuration
For image editing, configure similar settings but with an additional option:
- **Google API Key**: Your Google API key
- **Model**: Select the image generation model (same options as text-to-image)
- **Max References**: Configure the maximum number of reference images (1-3). This determines how many reference images you can provide when editing an image.
![A dark-themed configuration interface for the "Visualizer" application displaying settings for the "Google Image Editing" tab. The panel features a configured Google API key section and a dropdown menu selecting the "gemini-3-pro-image-preview" model. A slider control at the bottom sets the "Max References" value to 3.](/talemate/img/0.34.0/visual-agent-google-6.png)
!!! note "Reference Images"
Google's image editing models can use up to 3 reference images to guide the editing process. The "Max References" setting controls how many reference images Talemate will send to the API. You can adjust this based on your needs, but keep in mind that more references may provide better context for complex edits.
### Image Analysis Configuration
For image analysis, configure the following:
- **Google API Key**: Your Google API key
- **Model**: Select a vision-capable text model:
- **gemini-2.5-flash**: Fast analysis, good for general use
- **gemini-2.5-pro**: Higher quality analysis
- **gemini-3-pro-preview**: Latest model with improved capabilities
!!! note "Analysis Models"
Image analysis uses text models that support vision capabilities, not the image generation models. These models can analyze images and provide detailed descriptions, answer questions about image content, and extract information from visual content.
## Usage
Once configured, the Google backend will appear in the Visualizer agent status with green indicators showing which capabilities are available.
![A dark-themed user interface panel titled "Visualizer" marked with a green status indicator. Below the title are several clickable buttons, including a "References 3" button and four "Google" buttons distinguished by icons representing screen, image, edit, and search functions.](/talemate/img/0.34.0/visual-agent-google-8.png)
The status indicators show:
- **Text to Image**: Available when text-to-image backend is configured
- **Image Edit**: Available when image editing backend is configured (shows max references if configured)
- **Image Analysis**: Available when image analysis backend is configured
## Model Recommendations
### Text-to-Image and Image Editing
- **gemini-2.5-flash-image**: Best for faster generation and general use. Good balance of speed and quality.
- **gemini-3-pro-image-preview**: Best for higher quality results when speed is less important. Use when you need the best possible image quality.
### Image Analysis
- **gemini-2.5-flash**: Best for quick analysis and general use cases. Fast responses with good accuracy.
- **gemini-2.5-pro**: Best for detailed analysis requiring higher accuracy and more nuanced understanding.
- **gemini-3-pro-preview**: Best for the latest capabilities and most advanced analysis features.
## Prompt Formatting
The Google backend uses **Descriptive** prompt formatting by default. This means prompts are formatted as natural language descriptions rather than keyword lists. This works well with Google's Gemini models, which are designed to understand natural language instructions.
When generating images, provide detailed descriptions of what you want to create. For image editing, describe the changes you want to make in natural language.

View File

@@ -1,121 +0,0 @@
# OpenAI
The OpenAI backend provides image generation, editing, and analysis capabilities using OpenAI's image models. It supports text-to-image generation with DALL·E 3 and GPT-Image models, image editing with GPT-Image models, and AI-powered image analysis using vision-capable GPT models.
![The image displays the "General" settings tab of the "Visualizer" interface, featuring a dark-themed layout with a sidebar menu on the left. The main panel includes dropdown menus where "Google" is selected as the client and "OpenAI" is chosen for text-to-image, image editing, and image analysis backends. Additional controls show an image generation timeout slider set to 301, checkboxes for automatic setup and generation, and a selector for the fallback prompt type.](/talemate/img/0.34.0/visual-agent-openai-1.png)
## Prerequisites
Before configuring the OpenAI backend, you need to obtain an OpenAI API key:
1. Go to [OpenAI Platform](https://platform.openai.com/api-keys)
2. Sign in with your OpenAI account
3. Create a new API key or use an existing one
4. Copy the API key
Then configure it in Talemate:
1. Open Talemate Settings → Application → OpenAI API
2. Paste your OpenAI API key in the "OpenAI API Key" field
3. Save your changes
For additional instructions, see the [OpenAI API setup guide](/talemate/user-guide/apis/openai/).
## Configuration
In the Visualizer agent settings, select OpenAI as your backend for text-to-image generation, image editing, image analysis, or any combination of these. Each operation can be configured separately.
### Text-to-Image Configuration
For text-to-image generation, configure the following settings:
- **OpenAI API Key**: Your OpenAI API key (configured globally in Talemate Settings)
- **Model**: Select the image generation model to use:
- **dall-e-3**: OpenAI's DALL·E 3 model (widely available)
- **gpt-image-1**: OpenAI's GPT-Image model (may require organization verification)
- **gpt-image-1-mini**: Smaller version of GPT-Image (may require organization verification)
![A screenshot of the "Visualizer" application settings interface with the "OpenAI Text to Image" tab selected on the left sidebar. The main panel displays a masked input field for a configured OpenAI API key and a dropdown menu set to the "dall-e-3" model.](/talemate/img/0.34.0/visual-agent-openai-2.png)
!!! warning "Organization Verification"
The **gpt-image-1** and **gpt-image-1-mini** models may require your OpenAI organization to be verified before you can use them. If you encounter errors with these models, you may need to complete OpenAI's organization verification process.
!!! note "Model Testing Status"
Talemate's organization is not verified with OpenAI, and we have not tested the **gpt-image-1** and **gpt-image-1-mini** models. We have confirmed that **dall-e-3** works correctly. If you have access to the GPT-Image models and encounter issues, please report them so we can improve support for these models.
The OpenAI backend automatically sets resolution based on the format and model you select:
- **gpt-image-1** and **gpt-image-1-mini**:
- Landscape: 1536x1024
- Portrait: 1024x1536
- Square: 1024x1024
- **dall-e-3**:
- Landscape: 1792x1024
- Portrait: 1024x1792
- Square: 1024x1024
### Image Editing Configuration
For image editing, configure similar settings but note that DALL·E 3 does not support image editing:
- **OpenAI API Key**: Your OpenAI API key
- **Model**: Select an image editing model:
- **gpt-image-1**: Full-featured image editing model (may require organization verification)
- **gpt-image-1-mini**: Smaller image editing model (may require organization verification)
![This screenshot displays the settings interface for an application called "Visualizer," specifically showing the "OpenAI Image Editing" configuration panel. The right side features a dropdown menu for selecting the model "gpt-image-1" beneath a configured API key section. An orange notification box at the bottom alerts the user that this specific model may require OpenAI organization verification.](/talemate/img/0.34.0/visual-agent-openai-3.png)
!!! warning "DALL·E 3 Limitations"
DALL·E 3 does not support image editing. If you select DALL·E 3 for image editing, you will receive an error. Use **gpt-image-1** or **gpt-image-1-mini** for image editing instead.
!!! note "Reference Images"
OpenAI's image editing models support a single reference image. When editing an image, provide one reference image that will be used as the base for the edit.
### Image Analysis Configuration
For image analysis, configure the following:
- **OpenAI API Key**: Your OpenAI API key
- **Model**: Select a vision-capable text model:
- **gpt-4.1-mini**: Fast analysis model with vision capabilities
- **gpt-4o-mini**: Alternative vision model option
![This image shows the settings interface for an application named Visualizer, with the "OpenAI Image Analysis" tab selected on the left sidebar. The main panel allows users to configure the OpenAI vision API, displaying a confirmed API key status. A dropdown menu below specifically indicates that the "gpt-4.1-mini" model is selected.](/talemate/img/0.34.0/visual-agent-openai-4.png)
!!! note "Analysis Models"
Image analysis uses text models that support vision capabilities, not the image generation models. These models can analyze images and provide detailed descriptions, answer questions about image content, and extract information from visual content.
## Usage
Once configured, the OpenAI backend will appear in the Visualizer agent status with green indicators showing which capabilities are available.
![This image captures a dark-mode user interface section titled "Visualizer," marked by an active green status dot. Below the title, there are several pill-shaped tags or buttons representing data sources, including "Google," "References 1," and three distinct "OpenAI" options. The OpenAI buttons are highlighted in green, distinguishing them from the greyed-out Google and References buttons.](/talemate/img/0.34.0/visual-agent-openai-5.png)
The status indicators show:
- **Text to Image**: Available when text-to-image backend is configured
- **Image Edit**: Available when image editing backend is configured (shows "References 1" indicating single reference support)
- **Image Analysis**: Available when image analysis backend is configured
## Model Recommendations
### Text-to-Image
- **dall-e-3**: Most widely available option. Good for general use, though quality may vary.
- **gpt-image-1**: Higher quality option, but requires organization verification. Use if you have access and need better results.
- **gpt-image-1-mini**: Smaller version of GPT-Image, faster generation. Requires organization verification.
### Image Editing
- **gpt-image-1**: Best quality for image editing. Requires organization verification.
- **gpt-image-1-mini**: Faster editing option. Requires organization verification.
### Image Analysis
- **gpt-4.1-mini**: Recommended default for image analysis. Fast and accurate.
- **gpt-4o-mini**: Alternative option if you prefer this model.
## Prompt Formatting
The OpenAI backend uses **Descriptive** prompt formatting by default. This means prompts are formatted as natural language descriptions rather than keyword lists. Provide detailed, natural language descriptions of what you want to create or edit.

View File

@@ -1,119 +0,0 @@
# OpenRouter
The OpenRouter backend provides access to image generation, editing, and analysis capabilities through OpenRouter's unified API. OpenRouter allows you to access multiple AI providers through a single API, giving you flexibility to choose from various models and providers.
![A dark-themed settings interface for the "Visualizer" application, displaying a sidebar with General, OpenRouter, and Styles navigation options. The main panel allows configuration of backend services, showing "OpenRouter" selected for text-to-image, image editing, and image analysis, with "Google" set as the client. Additional controls include a slider for image generation timeout set to 301 and checkboxes for automatic setup and generation.](/talemate/img/0.34.0/visual-agent-openrouter-1.png)
## Prerequisites
Before configuring the OpenRouter backend, you need to obtain an OpenRouter API key:
1. Go to [OpenRouter Keys](https://openrouter.ai/settings/keys)
2. Sign in with your account
3. Create a new API key or use an existing one
4. Copy the API key
Then configure it in Talemate:
1. Open Talemate Settings → Application → OpenRouter API
2. Paste your OpenRouter API key in the "OpenRouter API Key" field
3. Save your changes
For additional instructions, see the [OpenRouter API setup guide](/talemate/user-guide/apis/openrouter/).
## Configuration
In the Visualizer agent settings, select OpenRouter as your backend for text-to-image generation, image editing, image analysis, or any combination of these. Each operation can be configured separately.
### Text-to-Image Configuration
For text-to-image generation, configure the following settings:
- **OpenRouter API Key**: Your OpenRouter API key (configured globally in Talemate Settings)
- **Model**: Select an image generation model from OpenRouter. The model list is dynamically populated based on models available through your OpenRouter account.
- **Only use these providers**: Optionally filter to specific providers (e.g., only use Google or OpenAI)
- **Ignore these providers**: Optionally exclude specific providers from consideration
![This screenshot depicts the "Visualizer" settings interface, specifically the "OpenRouter Text to Image" configuration tab. The panel displays an active API Key section, a model selection dropdown currently set to "google/gemini-2.5-flash-image", and additional options to filter specific service providers.](/talemate/img/0.34.0/visual-agent-openrouter-2.png)
!!! warning "Model Selection"
There is no reliable way for Talemate to determine which models support text-to-image generation, so the model list is unfiltered. Please consult the [OpenRouter documentation](https://openrouter.ai/docs) to verify that your selected model supports image generation before using it.
The OpenRouter backend automatically handles aspect ratios based on the format you select:
- **Landscape**: 16:9 aspect ratio
- **Portrait**: 9:16 aspect ratio
- **Square**: 1:1 aspect ratio
### Image Editing Configuration
For image editing, configure similar settings with an additional option:
- **OpenRouter API Key**: Your OpenRouter API key
- **Model**: Select an image editing model from OpenRouter
- **Max References**: Configure the maximum number of reference images (1-3). This determines how many reference images you can provide when editing an image.
- **Provider filtering**: Optionally filter providers (same as text-to-image)
![This screenshot displays the settings interface for an application named Visualizer, specifically focusing on the "OpenRouter - Image Editing" configuration tab. The main panel features input fields for an OpenRouter API key, a model selection dropdown set to "google/gemini-2.5-flash-image," and provider filtering options. Additionally, a slider at the bottom allows users to adjust the "Max References," which is currently set to 1.](/talemate/img/0.34.0/visual-agent-openrouter-3.png)
!!! warning "Model Selection"
There is no reliable way for Talemate to determine which models support image editing, so the model list is unfiltered. Image editing refers to image generation with support for 1 or more contextual reference images. Please consult the [OpenRouter documentation](https://openrouter.ai/docs) to verify that your selected model supports image editing before using it.
### Image Analysis Configuration
For image analysis, configure the following:
- **OpenRouter API Key**: Your OpenRouter API key
- **Model**: Select a vision-capable text model from OpenRouter
- **Provider filtering**: Optionally filter providers
![A screenshot of the "Visualizer" application interface showing the "OpenRouter Image Analysis" settings panel. The configuration area displays a model selection dropdown set to "google/gemini-2.5-flash" alongside a configured API key field. An informational box notes that the model list is unfiltered and users should verify that their chosen text generation model supports multi-modal vision capabilities.](/talemate/img/0.34.0/visual-agent-openrouter-4.png)
!!! warning "Model Selection"
There is no reliable way for Talemate to determine which models support image analysis, so the model list is unfiltered. Image analysis requires a text generation model that is multi-modal and supports vision capabilities. Please consult the [OpenRouter documentation](https://openrouter.ai/docs) to verify that your selected model supports vision before using it.
## Usage
Once configured, the OpenRouter backend will appear in the Visualizer agent status with green indicators showing which capabilities are available.
![A dark-mode user interface panel labeled "Visualizer" features a green status indicator dot next to the title. Below the header are several pill-shaped tags, including grey buttons for "Google" and "References 1" alongside three green "OpenRouter" buttons with various icons. This layout likely represents a configuration of active tools or API connections within a software application.](/talemate/img/0.34.0/visual-agent-openrouter-5.png)
The status indicators show:
- **Text to Image**: Available when text-to-image backend is configured
- **Image Edit**: Available when image editing backend is configured (shows max references if configured)
- **Image Analysis**: Available when image analysis backend is configured
## Model Recommendations
OpenRouter provides access to many models from different providers. Here are some general recommendations:
### Text-to-Image and Image Editing
- **google/gemini-2.5-flash-image**: Fast image generation with good quality
- **google/gemini-3-pro-image-preview**: Higher quality option (if available)
### Image Analysis
- **google/gemini-2.5-flash**: Fast analysis with good accuracy
- **google/gemini-2.5-pro**: Higher quality analysis
- **google/gemini-3-pro-preview**: Latest capabilities (if available)
## Provider Filtering
OpenRouter allows you to filter which providers are used for a specific model. This can be useful if:
- You want to use a specific provider for cost or quality reasons
- You want to avoid certain providers
- You want to test different providers for the same model
You can configure provider filtering in each backend's settings:
- **Only use these providers**: Limits requests to only the selected providers
- **Ignore these providers**: Excludes the selected providers from consideration
If both are configured, "Only use these providers" takes precedence.
## Prompt Formatting
The OpenRouter backend uses **Descriptive** prompt formatting by default. This means prompts are formatted as natural language descriptions rather than keyword lists. Provide detailed, natural language descriptions of what you want to create or edit.

View File

@@ -1,104 +0,0 @@
# SD.Next
The SD.Next backend provides image generation and editing capabilities using Stable Diffusion Next (SD.Next), a fork of AUTOMATIC1111's Stable Diffusion WebUI. SD.Next offers improved performance and additional features while maintaining compatibility with the AUTOMATIC1111 API.
![This screenshot displays the "General" settings menu of the "Visualizer" interface, featuring a dark theme with purple accents. Configuration options show "Google" selected as the client, with "SD.Next" set as the backend for both text-to-image and image editing tasks. The panel also includes an image generation timeout slider set to 301, a checked "Automatic Setup" box, and a "Fallback Prompt Type" dropdown set to Keywords.](/talemate/img/0.34.0/visual-agent-sdnext-1.png)
## Prerequisites
Before configuring the SD.Next backend, you need to have SD.Next installed and running. SD.Next can be run locally or accessed remotely via its API.
1. Install and start SD.Next on your system
2. Ensure the API is enabled and accessible
3. Note the API URL (default is `http://localhost:7860`)
## Configuration
In the Visualizer agent settings, select SD.Next as your backend for text-to-image generation, image editing, or both. You'll need to configure each backend separately if you want to use SD.Next for different operations.
### Text-to-Image Configuration
For text-to-image generation, configure the following settings:
- **API URL**: The URL where your SD.Next instance is running (e.g., `http://localhost:7860`)
- **Authentication Method**: Choose the authentication method:
- **None**: No authentication required
- **Basic (username/password)**: Use username and password authentication
- **Bearer (API Key)**: Use API key authentication
!!! note "ArliAI SD.Next Endpoints"
If you're connecting to ArliAI's SD.Next endpoints, you should use **Bearer (API Key)** authentication method. Configure your API key in the authentication settings.
- **Username/Password** (if using Basic auth): Your SD.Next credentials
- **API Key** (if using Bearer auth): Your API key for SD.Next
- **Steps**: Number of sampling steps (default: 40, range: 5-150)
- **Sampling Method**: The sampling algorithm to use (dynamically populated from your SD.Next instance)
- **CFG Scale**: Classifier-free guidance scale (default: 7.0, range: 1-30)
- **Model**: Select the model to use from your SD.Next models directory (dynamically populated)
- **Prompt Type**: Choose between "Keywords" or "Descriptive" prompt formatting
- **Resolutions**: Configure the pixel dimensions for Square, Portrait, and Landscape formats
![This screenshot displays the "Visualizer" interface for SD.Next text-to-image generation, featuring configuration settings for the API URL, 40 sampling steps, DPM++ 2M method, and a CFG scale of 7.0. The panel also shows the selected model as "juggernautXL_juggXlByRundiffusion" along with adjustable resolution presets for square, portrait, and landscape formats.](/talemate/img/0.34.0/visual-agent-sdnext-2.png)
![This screenshot displays a dark-themed user interface form with a dropdown menu labeled "Authentication Method" selected to "Basic (username/password)". Below the menu are two text input fields for entering a "Username" and "Password" under a section header labeled "AUTHENTICATION (OPTIONAL, SERVER DEPENDENT)".](/talemate/img/0.34.0/visual-agent-sdnext-3.png)
![This image displays a dark-themed user interface configuration panel for setting up authentication parameters. It features a dropdown menu where "Bearer (API Key)" is selected as the Authentication Method, positioned above a text input field labeled "API Key".](/talemate/img/0.34.0/visual-agent-sdnext-4.png)
### Image Editing Configuration
For image editing, configure similar settings. SD.Next supports image editing through its img2img API, which uses a single reference image.
![User interface for the "Visualizer" software displaying the "Image editing configuration for SD.Next" panel with a dark theme. It features adjustable sliders for Steps (set to 40) and CFG Scale (set to 7.0), alongside dropdown menus for the Sampling Method and Model selection. The bottom section includes input fields for defining specific pixel dimensions for square, portrait, and landscape image resolutions.](/talemate/img/0.34.0/visual-agent-sdnext-5.png)
!!! note "Reference Images"
SD.Next image editing supports a single reference image. When editing an image, provide one reference image that will be used as the base for the edit.
## Usage
Once configured, the SD.Next backend will appear in the Visualizer agent status with green indicators showing which capabilities are available.
![A dark mode user interface section titled "Visualizer," indicated by a green status dot. Below the header is a row of four buttons: "Google," two distinct "SD.Next" buttons with image and pencil icons respectively, and a "References 1" button.](/talemate/img/0.34.0/visual-agent-sdnext-6.png)
The status indicators show:
- **Text to Image**: Available when text-to-image backend is configured
- **Image Edit**: Available when image editing backend is configured (shows "References 1" indicating single reference support)
## Model and Sampler Selection
SD.Next dynamically fetches the list of available models and samplers from your instance when you configure the backend. This means:
- **Models**: The model dropdown is automatically populated with models available in your SD.Next installation
- **Samplers**: The sampling method dropdown is automatically populated with samplers available in your SD.Next instance
If you change the API URL or authentication settings, Talemate will automatically refresh the model and sampler lists from the new instance.
!!! tip "Model Selection"
If you don't select a specific model, SD.Next will use its default model. You can select "- Default Model -" from the dropdown to explicitly use the default, or leave the field empty.
## Sampler Settings
SD.Next provides extensive control over the generation process:
- **Steps**: More steps generally produce higher quality images but take longer. Typical values range from 20-50 steps, with 40 being a good default.
- **Sampling Method**: Different samplers produce different results. Popular options include:
- **DPM++ 2M**: Fast and high quality (default)
- **Euler a**: Fast, good for quick iterations
- **DPM++ 2M Karras**: Variant with different characteristics
- **CFG Scale**: Controls how closely the model follows your prompt. Lower values (1-7) allow more creative freedom, higher values (7-15) stick closer to the prompt.
## Prompt Formatting
SD.Next uses **Keywords** prompt formatting by default. This means prompts are formatted as keyword lists optimized for Stable Diffusion models. You can switch to **Descriptive** formatting if you prefer natural language descriptions, though Keywords typically work better with SD models.
## Remote Access
If you're running SD.Next on a remote server:
1. Configure SD.Next to listen on the appropriate network interface
2. Use the server's IP address or hostname in the API URL (e.g., `http://192.168.1.100:7860`)
3. Configure appropriate authentication if your SD.Next instance requires it
4. Ensure your firewall allows connections to the SD.Next port
!!! warning "Security Considerations"
If exposing SD.Next over a network, always use authentication. Unauthenticated SD.Next instances can be accessed by anyone on your network, which may pose security risks.

Some files were not shown because too many files have changed in this diff Show More